1
|
Cotner CE, O’Donnell E. Understanding the Landscape of Multi-Cancer Detection Tests: The Current Data and Clinical Considerations. Life (Basel) 2024; 14:896. [PMID: 39063649 PMCID: PMC11278188 DOI: 10.3390/life14070896] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Revised: 07/09/2024] [Accepted: 07/12/2024] [Indexed: 07/28/2024] Open
Abstract
Multi-cancer detection (MCD) tests are blood-based assays that screen for multiple cancers concurrently and offer a promising approach to improve early cancer detection and screening uptake. To date, there have been two prospective interventional studies evaluating MCD tests as a screening tool in human subjects. No MCD tests are currently approved by the FDA, but there is one commercially available MCD test. Ongoing trials continue to assess the efficacy, safety, and cost implications of MCD tests. In this review, we discuss the performance of CancerSEEK and Galleri, two leading MCD platforms, and discuss the clinical consideration for the broader application of this new technology.
Collapse
Affiliation(s)
- Cody E. Cotner
- Department of Medicine, Brigham and Women’s Hospital, Boston, MA 02115, USA;
- Harvard Medical School, Boston, MA 02115, USA
| | - Elizabeth O’Donnell
- Department of Medicine, Brigham and Women’s Hospital, Boston, MA 02115, USA;
- Harvard Medical School, Boston, MA 02115, USA
- Department of Medical Oncology, Dana-Farber Cancer Institute, 450 Brookline Ave. Boston, Boston, MA 02115, USA
| |
Collapse
|
2
|
Joskowicz L, Di Veroli B, Lederman R, Shoshan Y, Sosna J. Three scans are better than two for follow-up: An automatic method for finding missed and misidentified lesions in cross-sectional follow-up of oncology patients. Eur J Radiol 2024; 176:111530. [PMID: 38810439 DOI: 10.1016/j.ejrad.2024.111530] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2024] [Revised: 05/13/2024] [Accepted: 05/24/2024] [Indexed: 05/31/2024]
Abstract
PURPOSE Missed and misidentified neoplastic lesions in longitudinal studies of oncology patients are pervasive and may affect the evaluation of the disease status. Two newly identified patterns of lesion changes, lone lesions and non-consecutive lesion changes, may help radiologists to detect these lesions. This study evaluated a new interpretation revision workflow of lesion annotations in three or more consecutive scans based on these suspicious patterns. METHODS The interpretation revision workflow was evaluated on manual and computed lesion annotations in longitudinal oncology patient studies. For the manual revision, a senior radiologist and a senior neurosurgeon (the readers) manually annotated the lesions in each scan and later revised their annotations to identify missed and misidentified lesions with the workflow using the automatically detected patterns. For the computerized revision, lesion annotations were first computed with a previously trained nnU-Net and were then automatically revised with an AI-based method that automates the workflow readers' decisions. The evaluation included 67 patient studies with 2295 metastatic lesions in lung (19 patients, 83 CT scans, 1178 lesions), liver (18 patients, 77 CECT scans, 800 lesions) and brain (30 patients, 102 T1W-Gad MRI scans, 317 lesions). RESULTS Revision of the manual lesion annotations revealed 120 missed lesions and 20 misidentified lesions in 31 out of 67 (46%) studies. The automatic revision reduced the number of computed missed lesions by 55 and computed misidentified lesions by 164 in 51 out of 67 (76%) studies. CONCLUSION Automatic analysis of three or more consecutive volumetric scans helps find missed and misidentified lesions and may improve the evaluation of temporal changes of oncological lesions.
Collapse
Affiliation(s)
- Leo Joskowicz
- School of Computer Science and Engineering, The Hebrew University of Jerusalem, Israel.
| | - Beniamin Di Veroli
- School of Computer Science and Engineering, The Hebrew University of Jerusalem, Israel
| | - Richard Lederman
- Dept of Radiology, Hadassah Hebrew University Medical Center, Jerusalem, Israel
| | - Yigal Shoshan
- Dept of Neurosurgery, Hadassah Hebrew University Medical Center, Jerusalem, Israel
| | - Jacob Sosna
- Dept of Radiology, Hadassah Hebrew University Medical Center, Jerusalem, Israel
| |
Collapse
|
3
|
Dehghan Rouzi M, Moshiri B, Khoshnevisan M, Akhaee MA, Jaryani F, Salehi Nasab S, Lee M. Breast Cancer Detection with an Ensemble of Deep Learning Networks Using a Consensus-Adaptive Weighting Method. J Imaging 2023; 9:247. [PMID: 37998094 PMCID: PMC10671922 DOI: 10.3390/jimaging9110247] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2023] [Revised: 10/20/2023] [Accepted: 10/24/2023] [Indexed: 11/25/2023] Open
Abstract
Breast cancer's high mortality rate is often linked to late diagnosis, with mammograms as key but sometimes limited tools in early detection. To enhance diagnostic accuracy and speed, this study introduces a novel computer-aided detection (CAD) ensemble system. This system incorporates advanced deep learning networks-EfficientNet, Xception, MobileNetV2, InceptionV3, and Resnet50-integrated via our innovative consensus-adaptive weighting (CAW) method. This method permits the dynamic adjustment of multiple deep networks, bolstering the system's detection capabilities. Our approach also addresses a major challenge in pixel-level data annotation of faster R-CNNs, highlighted in a prominent previous study. Evaluations on various datasets, including the cropped DDSM (Digital Database for Screening Mammography), DDSM, and INbreast, demonstrated the system's superior performance. In particular, our CAD system showed marked improvement on the cropped DDSM dataset, enhancing detection rates by approximately 1.59% and achieving an accuracy of 95.48%. This innovative system represents a significant advancement in early breast cancer detection, offering the potential for more precise and timely diagnosis, ultimately fostering improved patient outcomes.
Collapse
Affiliation(s)
- Mohammad Dehghan Rouzi
- School of Electrical and computer Engineering, College of Engineering, University of Tehran, Tehran 14174-66191, Iran; (M.D.R.); (B.M.); (M.A.A.)
| | - Behzad Moshiri
- School of Electrical and computer Engineering, College of Engineering, University of Tehran, Tehran 14174-66191, Iran; (M.D.R.); (B.M.); (M.A.A.)
- Department of Electrical and Computer Engineering, University of Waterloo, Ontario, ON N2L 3G1, Canada
| | | | - Mohammad Ali Akhaee
- School of Electrical and computer Engineering, College of Engineering, University of Tehran, Tehran 14174-66191, Iran; (M.D.R.); (B.M.); (M.A.A.)
| | - Farhang Jaryani
- Human Genome Sequencing Center, Baylor College of Medicine, Houston, TX 77030, USA;
| | - Samaneh Salehi Nasab
- Department of Computer Engineering, Lorestan University, Khorramabad 68151-44316, Iran;
| | - Myeounggon Lee
- College of Health Sciences, Dong-A University, Saha-gu, Busan 49315, Republic of Korea
| |
Collapse
|
4
|
Ahn JS, Shin S, Yang SA, Park EK, Kim KH, Cho SI, Ock CY, Kim S. Artificial Intelligence in Breast Cancer Diagnosis and Personalized Medicine. J Breast Cancer 2023; 26:405-435. [PMID: 37926067 PMCID: PMC10625863 DOI: 10.4048/jbc.2023.26.e45] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2023] [Revised: 09/25/2023] [Accepted: 10/06/2023] [Indexed: 11/07/2023] Open
Abstract
Breast cancer is a significant cause of cancer-related mortality in women worldwide. Early and precise diagnosis is crucial, and clinical outcomes can be markedly enhanced. The rise of artificial intelligence (AI) has ushered in a new era, notably in image analysis, paving the way for major advancements in breast cancer diagnosis and individualized treatment regimens. In the diagnostic workflow for patients with breast cancer, the role of AI encompasses screening, diagnosis, staging, biomarker evaluation, prognostication, and therapeutic response prediction. Although its potential is immense, its complete integration into clinical practice is challenging. Particularly, these challenges include the imperatives for extensive clinical validation, model generalizability, navigating the "black-box" conundrum, and pragmatic considerations of embedding AI into everyday clinical environments. In this review, we comprehensively explored the diverse applications of AI in breast cancer care, underlining its transformative promise and existing impediments. In radiology, we specifically address AI in mammography, tomosynthesis, risk prediction models, and supplementary imaging methods, including magnetic resonance imaging and ultrasound. In pathology, our focus is on AI applications for pathologic diagnosis, evaluation of biomarkers, and predictions related to genetic alterations, treatment response, and prognosis in the context of breast cancer diagnosis and treatment. Our discussion underscores the transformative potential of AI in breast cancer management and emphasizes the importance of focused research to realize the full spectrum of benefits of AI in patient care.
Collapse
Affiliation(s)
| | | | | | | | | | | | | | - Seokhwi Kim
- Department of Pathology, Ajou University School of Medicine, Suwon, Korea
- Department of Biomedical Sciences, Ajou University Graduate School of Medicine, Suwon, Korea.
| |
Collapse
|
5
|
Liu J, Lei J, Ou Y, Zhao Y, Tuo X, Zhang B, Shen M. Mammography diagnosis of breast cancer screening through machine learning: a systematic review and meta-analysis. Clin Exp Med 2023; 23:2341-2356. [PMID: 36242643 DOI: 10.1007/s10238-022-00895-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2022] [Accepted: 09/12/2022] [Indexed: 12/24/2022]
Abstract
Breast cancer was the fourth leading cause of cancer-related death worldwide, and early mammography screening could decrease the breast cancer mortality. Artificial intelligence (AI)-assisted diagnose system based on machine learning (ML) methods can help improve the screening accuracy and efficacy. This study aimed to systematically review and make a meta-analysis on the diagnostic accuracy of mammography diagnosis of breast cancer through various ML methods. Springer Link, Science Direct (Elsevier), IEEE Xplore, PubMed and Web of Science were searched for relevant studies published from January 2000 to September 2021. The study was registered with the PROSPERO International Prospective Register of Systematic Reviews (protocol no. CRD42021284227). A Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) was used to assess the included studies, and reporting was evaluated using the Preferred Reporting Items for Systematic reviews and Meta-Analyses (PRISMA). The pooled summary estimates for sensitivity, specificity, the area under the receiver operating characteristic curve (AUC) for three ML methods (convolutional neural network [CNN], artificial neural network [ANN], support vector machine [SVM]) were calculated. A total of 32 studies with 23,804 images were included in the meta-analysis. The overall pooled estimate for sensitivity, specificity and AUC was 0.914 [95% CI 0.868-0.945], 0.916 [95% CI 0.873-0.945] and 0.945 for mammography diagnosis of breast cancer through three ML methods. The pooled sensitivity, specificity and AUC of CNN were 0.961 [95% CI 0.886-0.988], 0.950 [95% CI 0.924-0.967] and 0.974. The pooled sensitivity, specificity and AUC of ANN were 0.837 [95% CI 0.772-0.886], 0.894 [95% CI 0.764-0.957] and 0.881. The pooled sensitivity, specificity and AUC of SVM were 0.889 [95% CI 0.807-0.939], 0.843 [95% CI 0.724-0.916] and 0.913. Machine learning methods (especially CNN) show excellent performance in mammography diagnosis of breast cancer screening based on retrospective studies. More rigorous prospective studies are needed to evaluate the longitudinal performance of AI.
Collapse
Affiliation(s)
- Junjie Liu
- School of Medicine, Xi'an Jiaotong University, Xi'an, 710061, Shaanxi, People's Republic of China
| | - Jiangjie Lei
- School of Medicine, Xi'an Jiaotong University, Xi'an, 710061, Shaanxi, People's Republic of China
| | - Yuhang Ou
- School of Medicine, Xi'an Jiaotong University, Xi'an, 710061, Shaanxi, People's Republic of China
| | - Yilong Zhao
- School of Medicine, Xi'an Jiaotong University, Xi'an, 710061, Shaanxi, People's Republic of China
| | - Xiaofeng Tuo
- School of Medicine, Xi'an Jiaotong University, Xi'an, 710061, Shaanxi, People's Republic of China
| | - Baoming Zhang
- College of Stomatology, Xi'an Jiaotong University, Xi'an, 710004, Shaanxi, People's Republic of China
- Key laboratory of Shaanxi Province for Craniofacial Precision Medicine Research, College of Stomatology, Xi'an Jiaotong University, Xi'an, 710004, Shaanxi, People's Republic of China
- School of Public Health, Xi'an Jiaotong University Health Science Center, Xi'an, 710061, Shaanxi, People's Republic of China
| | - Mingwang Shen
- China-Australia Joint Research Center for Infectious Diseases, School of Public Health, Xi'an Jiaotong University Health Science Center, Xi'an, 710061, Shaanxi, People's Republic of China.
- Key Laboratory for Disease Prevention and Control and Health Promotion of Shaanxi Province, Xi'an Jiaotong University Health Science Center, Xi'an, 710061, Shaanxi, People's Republic of China.
| |
Collapse
|
6
|
Leming MJ, Bron EE, Bruffaerts R, Ou Y, Iglesias JE, Gollub RL, Im H. Challenges of implementing computer-aided diagnostic models for neuroimages in a clinical setting. NPJ Digit Med 2023; 6:129. [PMID: 37443276 DOI: 10.1038/s41746-023-00868-x] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2022] [Accepted: 06/26/2023] [Indexed: 07/15/2023] Open
Abstract
Advances in artificial intelligence have cultivated a strong interest in developing and validating the clinical utilities of computer-aided diagnostic models. Machine learning for diagnostic neuroimaging has often been applied to detect psychological and neurological disorders, typically on small-scale datasets or data collected in a research setting. With the collection and collation of an ever-growing number of public datasets that researchers can freely access, much work has been done in adapting machine learning models to classify these neuroimages by diseases such as Alzheimer's, ADHD, autism, bipolar disorder, and so on. These studies often come with the promise of being implemented clinically, but despite intense interest in this topic in the laboratory, limited progress has been made in clinical implementation. In this review, we analyze challenges specific to the clinical implementation of diagnostic AI models for neuroimaging data, looking at the differences between laboratory and clinical settings, the inherent limitations of diagnostic AI, and the different incentives and skill sets between research institutions, technology companies, and hospitals. These complexities need to be recognized in the translation of diagnostic AI for neuroimaging from the laboratory to the clinic.
Collapse
Affiliation(s)
- Matthew J Leming
- Center for Systems Biology, Massachusetts General Hospital, Boston, MA, USA.
- Massachusetts Alzheimer's Disease Research Center, Charlestown, MA, USA.
| | - Esther E Bron
- Department of Radiology and Nuclear Medicine, Erasmus MC, Rotterdam, The Netherlands
| | - Rose Bruffaerts
- Computational Neurology, Experimental Neurobiology Unit (ENU), Department of Biomedical Sciences, University of Antwerp, Antwerp, Belgium
- Biomedical Research Institute, Hasselt University, Diepenbeek, Belgium
| | - Yangming Ou
- Boston Children's Hospital, 300 Longwood Ave, Boston, MA, USA
| | - Juan Eugenio Iglesias
- Center for Medical Image Computing, University College London, London, UK
- Martinos Center for Biomedical Imaging, Harvard Medical School, Boston, MA, USA
- Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Randy L Gollub
- Department of Psychiatry, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Hyungsoon Im
- Center for Systems Biology, Massachusetts General Hospital, Boston, MA, USA.
- Massachusetts Alzheimer's Disease Research Center, Charlestown, MA, USA.
- Department of Radiology, Massachusetts General Hospital, Boston, MA, USA.
| |
Collapse
|
7
|
Kunar MA, Watson DG. Framing the fallibility of Computer-Aided Detection aids cancer detection. Cogn Res Princ Implic 2023; 8:30. [PMID: 37222932 PMCID: PMC10209366 DOI: 10.1186/s41235-023-00485-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2022] [Accepted: 04/29/2023] [Indexed: 05/25/2023] Open
Abstract
Computer-Aided Detection (CAD) has been proposed to help operators search for cancers in mammograms. Previous studies have found that although accurate CAD leads to an improvement in cancer detection, inaccurate CAD leads to an increase in both missed cancers and false alarms. This is known as the over-reliance effect. We investigated whether providing framing statements of CAD fallibility could keep the benefits of CAD while reducing over-reliance. In Experiment 1, participants were told about the benefits or costs of CAD, prior to the experiment. Experiment 2 was similar, except that participants were given a stronger warning and instruction set in relation to the costs of CAD. The results showed that although there was no effect of framing in Experiment 1, a stronger message in Experiment 2 led to a reduction in the over-reliance effect. A similar result was found in Experiment 3 where the target had a lower prevalence. The results show that although the presence of CAD can result in over-reliance on the technology, these effects can be mitigated by framing and instruction sets in relation to CAD fallibility.
Collapse
Affiliation(s)
- Melina A Kunar
- Department of Psychology, The University of Warwick, Coventry, CV4 7AL, UK.
| | - Derrick G Watson
- Department of Psychology, The University of Warwick, Coventry, CV4 7AL, UK
| |
Collapse
|
8
|
Letter H, Peratikos M, Toledano A, Hoffmeister J, Nishikawa R, Conant E, Shisler J, Maimone S, Diaz de Villegas H. Use of Artificial Intelligence for Digital Breast Tomosynthesis Screening: A Preliminary Real-world Experience. JOURNAL OF BREAST IMAGING 2023; 5:258-266. [PMID: 38416890 DOI: 10.1093/jbi/wbad015] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2022] [Indexed: 03/01/2024]
Abstract
OBJECTIVE The purpose of this study is to assess the "real-world" impact of an artificial intelligence (AI) tool designed to detect breast cancer in digital breast tomosynthesis (DBT) screening exams following 12 months of utilization in a subspecialized academic breast center. METHODS Following IRB approval, mammography audit reports, as specified in the BI-RADS atlas, were retrospectively generated for five radiologists reading at three locations during a 12-month time frame. One location had the AI tool (iCAD ProFound AI v2.0), and the other two locations did not. The co-primary endpoints were cancer detection rate (CDR) and abnormal interpretation rate (AIR). Secondary endpoints included positive predictive values (PPVs) for cancer among screenings with abnormal interpretations (PPV1) and for biopsies performed (PPV3). Odds ratios (OR) with two-sided 95% confidence intervals (CIs) summarized the impact of AI across radiologists using generalized estimating equations. RESULTS Nonsignificant differences were observed in CDR, AIR, and PPVs. The CDR was 7.3 with AI and 5.9 without AI (OR 1.3, 95% CI: 0.9-1.7). The AIR was 11.7% with AI and 11.8% without AI (OR 1.0, 95% CI: 0.8-1.3). The PPV1 was 6.2% with AI and 5.0% without AI (OR 1.3, 95% CI: 0.97-1.7). The PPV3 was 33.3% with AI and 32.0% without AI (OR 1.1, 95% CI: 0.8-1.5). CONCLUSION Although we are unable to show statistically significant changes in CDR and AIR outcomes in the two groups, the results are consistent with prior reader studies. There is a nonsignificant trend toward improvement in CDR with AI, without significant increases in AIR.
Collapse
Affiliation(s)
- Haley Letter
- Mayo Clinic, Department of Radiology, Jacksonville, FL, USA
- University of Florida, Department of Radiology, Jacksonville, FL, USA
| | | | | | | | - Robert Nishikawa
- University of Pittsburgh, Department of Radiology, Pittsburgh, PA, USA
| | - Emily Conant
- University of Pennsylvania, Department of Radiology, Philadelphia, PA, USA
| | | | - Santo Maimone
- Mayo Clinic, Department of Radiology, Jacksonville, FL, USA
| | | |
Collapse
|
9
|
Ng AY, Glocker B, Oberije C, Fox G, Sharma N, James JJ, Ambrózay É, Nash J, Karpati E, Kerruish S, Kecskemethy PD. Artificial Intelligence as Supporting Reader in Breast Screening: A Novel Workflow to Preserve Quality and Reduce Workload. JOURNAL OF BREAST IMAGING 2023; 5:267-276. [PMID: 38416889 DOI: 10.1093/jbi/wbad010] [Citation(s) in RCA: 11] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2022] [Indexed: 03/01/2024]
Abstract
OBJECTIVE To evaluate the effectiveness of a new strategy for using artificial intelligence (AI) as supporting reader for the detection of breast cancer in mammography-based double reading screening practice. METHODS Large-scale multi-site, multi-vendor data were used to retrospectively evaluate a new paradigm of AI-supported reading. Here, the AI served as the second reader only if it agrees with the recall/no-recall decision of the first human reader. Otherwise, a second human reader made an assessment followed by the standard clinical workflow. The data included 280 594 cases from 180 542 female participants screened for breast cancer at seven screening sites in two countries and using equipment from four hardware vendors. The statistical analysis included non-inferiority and superiority testing of cancer screening performance and evaluation of the reduction in workload, measured as arbitration rate and number of cases requiring second human reading. RESULTS Artificial intelligence as a supporting reader was found to be superior or noninferior on all screening metrics compared with human double reading while reducing the number of cases requiring second human reading by up to 87% (245 395/280 594). Compared with AI as an independent reader, the number of cases referred to arbitration was reduced from 13% (35 199/280 594) to 2% (5056/280 594). CONCLUSION The simulation indicates that the proposed workflow retains screening performance of human double reading while substantially reducing the workload. Further research should study the impact on the second human reader because they would only assess cases in which the AI prediction and first human reader disagree.
Collapse
Affiliation(s)
- Annie Y Ng
- Kheiron Medical Technologies, London, UK
| | - Ben Glocker
- Kheiron Medical Technologies, London, UK
- Imperial College London, Department of Computing, London, UK
| | | | | | - Nisha Sharma
- Leeds Teaching Hospital NHS Trust, Department of Radiology, Leeds, UK
| | - Jonathan J James
- Nottingham University Hospitals NHS Trust, Nottingham Breast Institute, Nottingham, UK
| | - Éva Ambrózay
- MaMMa Egészségügyi Zrt., Breast Diagnostic Department, Kecskemét, Hungary
| | | | | | | | | |
Collapse
|
10
|
Sharma N, Ng AY, James JJ, Khara G, Ambrózay É, Austin CC, Forrai G, Fox G, Glocker B, Heindl A, Karpati E, Rijken TM, Venkataraman V, Yearsley JE, Kecskemethy PD. Multi-vendor evaluation of artificial intelligence as an independent reader for double reading in breast cancer screening on 275,900 mammograms. BMC Cancer 2023; 23:460. [PMID: 37208717 DOI: 10.1186/s12885-023-10890-7] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2022] [Accepted: 04/26/2023] [Indexed: 05/21/2023] Open
Abstract
BACKGROUND Double reading (DR) in screening mammography increases cancer detection and lowers recall rates, but has sustainability challenges due to workforce shortages. Artificial intelligence (AI) as an independent reader (IR) in DR may provide a cost-effective solution with the potential to improve screening performance. Evidence for AI to generalise across different patient populations, screening programmes and equipment vendors, however, is still lacking. METHODS This retrospective study simulated DR with AI as an IR, using data representative of real-world deployments (275,900 cases, 177,882 participants) from four mammography equipment vendors, seven screening sites, and two countries. Non-inferiority and superiority were assessed for relevant screening metrics. RESULTS DR with AI, compared with human DR, showed at least non-inferior recall rate, cancer detection rate, sensitivity, specificity and positive predictive value (PPV) for each mammography vendor and site, and superior recall rate, specificity, and PPV for some. The simulation indicates that using AI would have increased arbitration rate (3.3% to 12.3%), but could have reduced human workload by 30.0% to 44.8%. CONCLUSIONS AI has potential as an IR in the DR workflow across different screening programmes, mammography equipment and geographies, substantially reducing human reader workload while maintaining or improving standard of care. TRIAL REGISTRATION ISRCTN18056078 (20/03/2019; retrospectively registered).
Collapse
Affiliation(s)
- Nisha Sharma
- The Leeds Teaching Hospital NHS Trust, Leeds, UK
| | - Annie Y Ng
- Kheiron Medical Technologies, London, UK.
| | - Jonathan J James
- Nottingham Breast Institute, City Hospital, Nottingham University Hospitals NHS Trust, Nottingham, UK
| | | | | | | | - Gábor Forrai
- Duna Medical Center, Budapest, Hungary
- GÉ-RAD Kft, Budapest, Hungary
| | | | - Ben Glocker
- Kheiron Medical Technologies, London, UK
- Department of Computing, Imperial College London, London, UK
| | | | - Edit Karpati
- Kheiron Medical Technologies, London, UK
- Medicover, Budapest, Hungary
| | | | | | | | | |
Collapse
|
11
|
Cushnan D, Young KC, Ward D, Halling-Brown MD, Duffy S, Given-Wilson R, Wallis MG, Wilkinson L, Lyburn I, Sidebottom R, McAvinchey R, Lewis EB, Mackenzie A, Warren LM. Lessons learned from independent external validation of an AI tool to detect breast cancer using a representative UK data set. Br J Radiol 2023; 96:20211104. [PMID: 36607283 PMCID: PMC9975375 DOI: 10.1259/bjr.20211104] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2021] [Revised: 11/21/2022] [Accepted: 11/30/2022] [Indexed: 01/07/2023] Open
Abstract
OBJECTIVE To pilot a process for the independent external validation of an artificial intelligence (AI) tool to detect breast cancer using data from the NHS breast screening programme (NHSBSP). METHODS A representative data set of mammography images from 26,000 women attending 2 NHS screening centres, and an enriched data set of 2054 positive cases were used from the OPTIMAM image database. The use case of the AI tool was the replacement of the first or second human reader. The performance of the AI tool was compared to that of human readers in the NHSBSP. RESULTS Recommendations for future external validations of AI tools to detect breast cancer are provided. The tool recalled different breast cancers to the human readers. This study showed the importance of testing AI tools on all types of cases (including non-standard) and the clarity of any warning messages. The acceptable difference in sensitivity and specificity between the AI tool and human readers should be determined. Any information vital for the clinical application should be a required output for the AI tool. It is recommended that the interaction of radiologists with the AI tool, and the effect of the AI tool on arbitration be investigated prior to clinical use. CONCLUSION This pilot demonstrated several lessons for future independent external validation of AI tools for breast cancer detection. ADVANCES IN KNOWLEDGE Knowledge has been gained towards best practice procedures for performing independent external validations of AI tools for the detection of breast cancer using data from the NHS Breast Screening Programme.
Collapse
Affiliation(s)
| | | | - Dominic Ward
- Royal Surrey NHS Foundation Trust, Guildford, United Kingdom
| | | | - Stephen Duffy
- Queen Mary University London, London, United Kingdom
| | | | - Matthew G Wallis
- Cambridge Breast Unit and NIHR Cambridge Biomedical Research Centre, Cambridge University Hospitals NHS Trust, Cambridge, United Kingdom
| | - Louise Wilkinson
- Oxford University Hospitals NHS Foundation Trust, Oxford, United Kingdom
| | | | | | | | - Emma B Lewis
- Royal Surrey NHS Foundation Trust, Guildford, United Kingdom
| | | | - Lucy M Warren
- Royal Surrey NHS Foundation Trust, Guildford, United Kingdom
| |
Collapse
|
12
|
Afriyie Y, Weyori BA, Opoku AA. A scaling up approach: a research agenda for medical imaging analysis with applications in deep learning. J EXP THEOR ARTIF IN 2023. [DOI: 10.1080/0952813x.2023.2165721] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/26/2023]
Affiliation(s)
- Yaw Afriyie
- Department of Computer Science and Informatics, University of Energy and Natural Resources, School of Sciences, Sunyani, Ghana
- Department of Computer Science, Faculty of Information and Communication Technology, SD Dombo University of Business and Integrated Development Studies, Wa, Ghana
| | - Benjamin A. Weyori
- Department of Computer Science and Informatics, University of Energy and Natural Resources, School of Sciences, Sunyani, Ghana
| | - Alex A. Opoku
- Department of Mathematics & Statistics, University of Energy and Natural Resources, School of Sciences, Sunyani, Ghana
| |
Collapse
|
13
|
Nicol ED, Weir-McCall JR, Shaw LJ, Williamson E. Great debates in cardiac computed tomography: OPINION: "Artificial intelligence and the future of cardiovascular CT - Managing expectation and challenging hype". J Cardiovasc Comput Tomogr 2023; 17:11-17. [PMID: 35977872 DOI: 10.1016/j.jcct.2022.07.005] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/01/2022] [Revised: 06/30/2022] [Accepted: 07/16/2022] [Indexed: 10/17/2022]
Abstract
This manuscript has been written as a follow-up to the "AI/ML great debate" featured at the 2021 Society of Cardiovascular Computed Tomography (SCCT) Annual Scientific Meeting. In debate style, we highlighti the need for expectation management of AI/ML, debunking the hype around current AI techniques, and countering the argument that in its current day format AI/ML is the "silver bullet" for the interpretation of daily clinical CCTA practice.
Collapse
Affiliation(s)
- Edward D Nicol
- Departments of Cardiology and Radiology, Royal Brompton Hospital, Guys and St Thomas' NHS Foundation Trust, London, UK; School of Biomedical Engineering and Imaging Sciences, King's College, London, UK.
| | - Jonathan R Weir-McCall
- School of Clinical Medicine, University of Cambridge, Cambridge, UK; Department of Radiology, Royal Papworth Hospital, Cambridge, UK
| | - Leslee J Shaw
- The Mount Sinai Hospital, 1468 Madison Ave, New York, NY 10029, United States
| | | |
Collapse
|
14
|
Chen Y, James JJ, Michalopoulou E, Darker IT, Jenkins J. Performance of Radiologists and Radiographers in Double Reading Mammograms: The UK National Health Service Breast Screening Program. Radiology 2023; 306:102-109. [PMID: 36098643 DOI: 10.1148/radiol.212951] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
Background Double reading can be used in screening mammography, but it is labor intensive. There is limited evidence on whether trained radiographers (ie, technologists) may be used to provide double reading. Purpose To compare the performance of radiologists and radiographers double reading screening mammograms, considering reader experience level. Materials and Methods In this retrospective study, performance and experience data were obtained for radiologists and radiographer readers of all screening mammograms in England from April 2015 to March 2016. Cancer detection rate (CDR), recall rate (RR), and positive predictive value (PPV) of recall based on biopsy-proven findings were calculated for first readers. Performance metrics were analyzed according to reader professional group and years of reading experience using the analysis of variance test. P values less than .05 were considered to indicate statistically significant difference. Results During the study period, 401 readers (224 radiologists and 177 radiographers) double read 1 404 395 screening digital mammograms. There was no difference in CDR between radiologist and radiographer readers (mean, 7.84 vs 7.53 per 1000 examinations, respectively; P = .08) and no difference for readers with more than 10 years of experience compared with 5 years or fewer years of experience, regardless of professional group (mean, 7.75 vs 7.71 per 1000 examinations respectively, P = .87). No difference in the mean RR was observed between radiologists and radiographer readers (5.0% vs 5.2%, respectively, P = .63). A lower RR was seen for readers with more than 10 years of experience compared with 5 years or fewer, regardless of professional group (mean, 4.8% vs 5.8%, respectively; P = .001). No variation in PPV was observed between them (P = .42), with PPV values of 17.1% for radiologists versus 16.1% for radiographers. A higher PPV was seen for readers with more than 10 years of experience compared with 5 years or less, regardless of professional group (mean, 17.5% and 14.9%, respectively; P = .02). Conclusion No difference in performance was observed between radiographers and radiologists reading screening mammograms in a program that used double reading. Published under a CC BY 4.0 license Online supplemental material is available for this article. See also the editorial by Hooley and Durand in this issue.
Collapse
Affiliation(s)
- Yan Chen
- From the University of Nottingham, School of Medicine, Division of Cancer and Stem Cells, City Hospital Campus, Hucknall Road, Nottingham, NG5 1PB, United Kingdom (Y.C., E.M., I.T.D.); Nottingham University Hospitals NHS Trust, Nottingham Breast Institute, City Hospital Campus, Nottingham, United Kingdom (J.J.J.); and Public Health Commissioning and Operations, Directorate of the Chief Operating Officer, NHS England and NHS Improvement, Redditch, United Kingdom (J.J.)
| | - Jonathan J James
- From the University of Nottingham, School of Medicine, Division of Cancer and Stem Cells, City Hospital Campus, Hucknall Road, Nottingham, NG5 1PB, United Kingdom (Y.C., E.M., I.T.D.); Nottingham University Hospitals NHS Trust, Nottingham Breast Institute, City Hospital Campus, Nottingham, United Kingdom (J.J.J.); and Public Health Commissioning and Operations, Directorate of the Chief Operating Officer, NHS England and NHS Improvement, Redditch, United Kingdom (J.J.)
| | - Eleni Michalopoulou
- From the University of Nottingham, School of Medicine, Division of Cancer and Stem Cells, City Hospital Campus, Hucknall Road, Nottingham, NG5 1PB, United Kingdom (Y.C., E.M., I.T.D.); Nottingham University Hospitals NHS Trust, Nottingham Breast Institute, City Hospital Campus, Nottingham, United Kingdom (J.J.J.); and Public Health Commissioning and Operations, Directorate of the Chief Operating Officer, NHS England and NHS Improvement, Redditch, United Kingdom (J.J.)
| | - Iain T Darker
- From the University of Nottingham, School of Medicine, Division of Cancer and Stem Cells, City Hospital Campus, Hucknall Road, Nottingham, NG5 1PB, United Kingdom (Y.C., E.M., I.T.D.); Nottingham University Hospitals NHS Trust, Nottingham Breast Institute, City Hospital Campus, Nottingham, United Kingdom (J.J.J.); and Public Health Commissioning and Operations, Directorate of the Chief Operating Officer, NHS England and NHS Improvement, Redditch, United Kingdom (J.J.)
| | - Jacquie Jenkins
- From the University of Nottingham, School of Medicine, Division of Cancer and Stem Cells, City Hospital Campus, Hucknall Road, Nottingham, NG5 1PB, United Kingdom (Y.C., E.M., I.T.D.); Nottingham University Hospitals NHS Trust, Nottingham Breast Institute, City Hospital Campus, Nottingham, United Kingdom (J.J.J.); and Public Health Commissioning and Operations, Directorate of the Chief Operating Officer, NHS England and NHS Improvement, Redditch, United Kingdom (J.J.)
| |
Collapse
|
15
|
Syed AH, Khan T. Evolution of research trends in artificial intelligence for breast cancer diagnosis and prognosis over the past two decades: A bibliometric analysis. Front Oncol 2022; 12:854927. [PMID: 36267967 PMCID: PMC9578338 DOI: 10.3389/fonc.2022.854927] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2022] [Accepted: 08/30/2022] [Indexed: 01/27/2023] Open
Abstract
Objective In recent years, among the available tools, the concurrent application of Artificial Intelligence (AI) has improved the diagnostic performance of breast cancer screening. In this context, the present study intends to provide a comprehensive overview of the evolution of AI for breast cancer diagnosis and prognosis research using bibliometric analysis. Methodology Therefore, in the present study, relevant peer-reviewed research articles published from 2000 to 2021 were downloaded from the Scopus and Web of Science (WOS) databases and later quantitatively analyzed and visualized using Bibliometrix (R package). Finally, open challenges areas were identified for future research work. Results The present study revealed that the number of literature studies published in AI for breast cancer detection and survival prediction has increased from 12 to 546 between the years 2000 to 2021. The United States of America (USA), the Republic of China, and India are the most productive publication-wise in this field. Furthermore, the USA leads in terms of the total citations; however, hungry and Holland take the lead positions in average citations per year. Wang J is the most productive author, and Zhan J is the most relevant author in this field. Stanford University in the USA is the most relevant affiliation by the number of published articles. The top 10 most relevant sources are Q1 journals with PLOS ONE and computer in Biology and Medicine are the leading journals in this field. The most trending topics related to our study, transfer learning and deep learning, were identified. Conclusion The present findings provide insight and research directions for policymakers and academic researchers for future collaboration and research in AI for breast cancer patients.
Collapse
Affiliation(s)
- Asif Hassan Syed
- Department of Computer Science, Faculty of Computing and Information Technology Rabigh (FCITR), King Abdulaziz University, Jeddah, Saudi Arabia
| | - Tabrej Khan
- Department of Information Systems, Faculty of Computing and Information Technology Rabigh (FCITR), King Abdulaziz University, Jeddah, Saudi Arabia
| |
Collapse
|
16
|
Lamb LR, Lehman CD, Gastounioti A, Conant EF, Bahl M. Artificial Intelligence (AI) for Screening Mammography, From the AJR Special Series on AI Applications. AJR Am J Roentgenol 2022; 219:369-380. [PMID: 35018795 DOI: 10.2214/ajr.21.27071] [Citation(s) in RCA: 22] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Artificial intelligence (AI) applications for screening mammography are being marketed for clinical use in the interpretative domains of lesion detection and diagnosis, triage, and breast density assessment and in the noninterpretive domains of breast cancer risk assessment, image quality control, image acquisition, and dose reduction. Evidence in support of these nascent applications, particularly for lesion detection and diagnosis, is largely based on multireader studies with cancer-enriched datasets rather than rigorous clinical evaluation aligned with the application's specific intended clinical use. This article reviews commercial AI algorithms for screening mammography that are currently available for clinical practice, their use, and evidence supporting their performance. Clinical implementation considerations, such as workflow integration, governance, and ethical issues, are also described. In addition, the future of AI for screening mammography is discussed, including the development of interpretive and noninterpretive AI applications and strategic priorities for research and development.
Collapse
Affiliation(s)
- Leslie R Lamb
- Department of Radiology, Massachusetts General Hospital, 55 Fruit St, WAC 240, Boston, MA 02114
| | - Constance D Lehman
- Department of Radiology, Massachusetts General Hospital, 55 Fruit St, WAC 240, Boston, MA 02114
| | - Aimilia Gastounioti
- Department of Radiology, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA
- Present affiliation: Mallinckrodt Institute of Radiology, Washington University School of Medicine, St. Louis, MO
| | - Emily F Conant
- Department of Radiology, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA
| | - Manisha Bahl
- Department of Radiology, Massachusetts General Hospital, 55 Fruit St, WAC 240, Boston, MA 02114
| |
Collapse
|
17
|
Affiliation(s)
- Thomas Penzel
- Corresponding author. Thomas Penzel, Interdisciplinary Sleep Medicine Center, Charite center for Pneumology CC12, Charite University Hospital, Chariteplatz 1, 10117 Berlin, Germany.
| |
Collapse
|
18
|
CoroNet: Deep Neural Network-Based End-to-End Training for Breast Cancer Diagnosis. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12147080] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/16/2022]
Abstract
In 2020, according to the publications of both the Global Cancer Observatory (GCO) and the World Health Organization (WHO), breast cancer (BC) represents one of the highest prevalent cancers in women worldwide. Almost 47% of the world’s 100,000 people are diagnosed with breast cancer, among females. Moreover, BC prevails among 38.8% of Egyptian women having cancer. Current deep learning developments have shown the common usage of deep convolutional neural networks (CNNs) for analyzing medical images. Unlike the randomly initialized ones, pre-trained natural image database (ImageNet)-based CNN models may become successfully fine-tuned to obtain improved findings. To conduct the automatic detection of BC by the CBIS-DDSM dataset, a CNN model, namely CoroNet, is proposed. It relies on the Xception architecture, which has been pre-trained on the ImageNet dataset and has been fully trained on whole-image BC according to mammograms. The convolutional design method is used in this paper, since it performs better than the other methods. On the prepared dataset, CoroNet was trained and tested. Experiments show that in a four-class classification, it may attain an overall accuracy of 94.92% (benign mass vs. malignant mass) and (benign calcification vs. malignant calcification). CoroNet has a classification accuracy of 88.67% for the two-class cases (calcifications and masses). The paper concluded that there are promising outcomes that could be improved because more training data are available.
Collapse
|
19
|
Forrai G, Kovács E, Ambrózay É, Barta M, Borbély K, Lengyel Z, Ormándi K, Péntek Z, Tünde T, Sebő É. Use of Diagnostic Imaging Modalities in Modern Screening, Diagnostics and Management of Breast Tumours 1st Central-Eastern European Professional Consensus Statement on Breast Cancer. Pathol Oncol Res 2022; 28:1610382. [PMID: 35755417 PMCID: PMC9214693 DOI: 10.3389/pore.2022.1610382] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/20/2022] [Accepted: 04/29/2022] [Indexed: 12/11/2022]
Abstract
Breast radiologists and nuclear medicine specialists updated their previous recommendation/guidance at the 4th Hungarian Breast Cancer Consensus Conference in Kecskemét. A recommendation is hereby made that breast tumours should be screened, diagnosed and treated according to these guidelines. These professional guidelines include the latest technical developments and research findings, including the role of imaging methods in therapy and follow-up. It includes details on domestic development proposals and also addresses related areas (forensic medicine, media, regulations, reimbursement). The entire material has been agreed with the related medical disciplines.
Collapse
Affiliation(s)
- Gábor Forrai
- GÉ-RAD Kft., Budapest, Hungary
- Duna Medical Center, Budapest, Hungary
| | - Eszter Kovács
- GÉ-RAD Kft., Budapest, Hungary
- Duna Medical Center, Budapest, Hungary
| | | | | | - Katalin Borbély
- National Institute of Oncology, Budapest, Hungary
- Ministry of Human Capacities, Budapest, Hungary
| | | | | | | | - Tasnádi Tünde
- Dr Réthy Pál Member Hospital of Békés County Central Hospital, Békéscsaba, Hungary
| | - Éva Sebő
- Kenézy Gyula University Hospital, University of Debrecen, Debrecen, Hungary
| |
Collapse
|
20
|
Spadaccini M, Hassan C, Alfarone L, Da Rio L, Maselli R, Carrara S, Galtieri PA, Pellegatta G, Fugazza A, Koleth G, Emmanuel J, Anderloni A, Mori Y, Wallace MB, Sharma P, Repici A. Comparing the number and relevance of false activations between 2 artificial intelligence computer-aided detection systems: the NOISE study. Gastrointest Endosc 2022; 95:975-981.e1. [PMID: 34995639 DOI: 10.1016/j.gie.2021.12.031] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/29/2021] [Accepted: 12/25/2021] [Indexed: 12/11/2022]
Abstract
BACKGROUND AND AIMS Artificial intelligence has been shown to be effective in polyp detection, and multiple computer-aided detection (CADe) systems have been developed. False-positive (FP) activation emerged as a possible way to benchmark CADe performance in clinical practice. The aim of this study was to validate a previously developed classification of FPs comparing the performances of different brands of approved CADe systems. METHODS We compared 2 different consecutive video libraries (40 video per arm) collected at Humanitas Research Hospital with 2 different CADe system brands (CADe A and CADe B). For each video, the number of CADe false activations, cause, and time spent by the endoscopist to examine the area erroneously highlighted were reported. The FP activations were classified according to the previously developed classification of FPs (the NOISE classification) according to their cause and relevance. RESULTS In CADe A 1021 FP activations were registered across the 40 videos (25.5 ± 12.2 FPs per colonoscopy), whereas in CADe B 1028 were identified (25.7 ± 13.2 FPs per colonoscopy; P = .53). Among them, 22.9 ± 9.9 (89.8% in CADe A) and 22.1 ± 10.0 (86.0% in CADe B) were because of artifacts from the bowel wall. Conversely, 2.6 ± 1.9 (10.2% in CADe A) and 3.5 ± 2.1 (14% in CADe B) were caused by bowel content (P = .45). Within CADe A each false activation required .2 ± .9 seconds, with 1.6 ± 1.0 FPs (6.3%) requiring additional time for endoscopic assessment. Comparable results were reported within CADe B with .2 ± .8 seconds spent per false activation and 1.8 ± 1.2 FPs per colonoscopy requiring additional inspection. CONCLUSIONS The use of a standardized nomenclature provided comparable results with either of the 2 recently approved CADe systems. (Clinical trial registration number: NCT04399590.).
Collapse
Affiliation(s)
- Marco Spadaccini
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy; Endoscopy Unit, Humanitas Clinical and Research Center-IRCCS, Rozzano, Italy
| | - Cesare Hassan
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy; Endoscopy Unit, Humanitas Clinical and Research Center-IRCCS, Rozzano, Italy
| | - Ludovico Alfarone
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy; Endoscopy Unit, Humanitas Clinical and Research Center-IRCCS, Rozzano, Italy
| | - Leonardo Da Rio
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy; Endoscopy Unit, Humanitas Clinical and Research Center-IRCCS, Rozzano, Italy
| | - Roberta Maselli
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy; Endoscopy Unit, Humanitas Clinical and Research Center-IRCCS, Rozzano, Italy
| | - Silvia Carrara
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy
| | | | - Gaia Pellegatta
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy
| | - Alessandro Fugazza
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy
| | - Glenn Koleth
- Department of Gastroenterology and Hepatology, Hospital Selayang, Selangor, Malaysia
| | - James Emmanuel
- Department of Gastroenterology and Hepatology, Queen Elizabeth Hospital, Sabah, Malaysia
| | - Andrea Anderloni
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy
| | - Yuichi Mori
- Clinical Effectiveness Research Group, Institute of Health and Society, Faculty of Medicine, University of Oslo, Oslo, Norway; Digestive Disease Center, Showa University Northern Yokohama Hospital, Yokohama, Japan
| | - Michael B Wallace
- Endoscopy Unit, Sheikh Shakhbout Medical City, Abu Dhabi, United Arab Emirates
| | - Prateek Sharma
- Department of Gastroenterology and Hepatology, Kansas City VA Medical Center, Kansas City, USA
| | - Alessandro Repici
- Department of Biomedical Sciences, Humanitas University, Pieve Emanuele, Italy; Endoscopy Unit, Humanitas Clinical and Research Center-IRCCS, Rozzano, Italy
| |
Collapse
|
21
|
Ha R, Jairam MP. A review of artificial intelligence in mammography. Clin Imaging 2022; 88:36-44. [DOI: 10.1016/j.clinimag.2022.05.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2021] [Revised: 04/28/2022] [Accepted: 05/12/2022] [Indexed: 11/16/2022]
|
22
|
Ueda D, Yamamoto A, Onoda N, Takashima T, Noda S, Kashiwagi S, Morisaki T, Fukumoto S, Shiba M, Morimura M, Shimono T, Kageyama K, Tatekawa H, Murai K, Honjo T, Shimazaki A, Kabata D, Miki Y. Development and validation of a deep learning model for detection of breast cancers in mammography from multi-institutional datasets. PLoS One 2022; 17:e0265751. [PMID: 35324962 PMCID: PMC8947392 DOI: 10.1371/journal.pone.0265751] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2021] [Accepted: 03/07/2022] [Indexed: 12/24/2022] Open
Abstract
Objectives The objective of this study was to develop and validate a state-of-the-art, deep learning (DL)-based model for detecting breast cancers on mammography. Methods Mammograms in a hospital development dataset, a hospital test dataset, and a clinic test dataset were retrospectively collected from January 2006 through December 2017 in Osaka City University Hospital and Medcity21 Clinic. The hospital development dataset and a publicly available digital database for screening mammography (DDSM) dataset were used to train and to validate the RetinaNet, one type of DL-based model, with five-fold cross-validation. The model’s sensitivity and mean false positive indications per image (mFPI) and partial area under the curve (AUC) with 1.0 mFPI for both test datasets were externally assessed with the test datasets. Results The hospital development dataset, hospital test dataset, clinic test dataset, and DDSM development dataset included a total of 3179 images (1448 malignant images), 491 images (225 malignant images), 2821 images (37 malignant images), and 1457 malignant images, respectively. The proposed model detected all cancers with a 0.45–0.47 mFPI and had partial AUCs of 0.93 in both test datasets. Conclusions The DL-based model developed for this study was able to detect all breast cancers with a very low mFPI. Our DL-based model achieved the highest performance to date, which might lead to improved diagnosis for breast cancer.
Collapse
Affiliation(s)
- Daiju Ueda
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka City University, Osaka, Japan
- * E-mail:
| | - Akira Yamamoto
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Naoyoshi Onoda
- Department of Breast and Endocrine Surgery, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Tsutomu Takashima
- Department of Breast and Endocrine Surgery, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Satoru Noda
- Department of Breast and Endocrine Surgery, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Shinichiro Kashiwagi
- Department of Breast and Endocrine Surgery, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Tamami Morisaki
- Department of Breast and Endocrine Surgery, Graduate School of Medicine, Osaka City University, Osaka, Japan
- Department of Premier Preventive Medicine, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Shinya Fukumoto
- Department of Premier Preventive Medicine, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Masatsugu Shiba
- Department of Gastroenterology, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Mina Morimura
- Department of General Practice, Osaka City University Hospital, Osaka, Japan
| | - Taro Shimono
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Ken Kageyama
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Hiroyuki Tatekawa
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Kazuki Murai
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Takashi Honjo
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Akitoshi Shimazaki
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Daijiro Kabata
- Department of Medical Statistics, Graduate School of Medicine, Osaka City University, Osaka, Japan
| | - Yukio Miki
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka City University, Osaka, Japan
| |
Collapse
|
23
|
Zhao W, Kang Q, Qian F, Li K, Zhu J, Ma B. Convolutional Neural Network-Based Computer-Assisted Diagnosis of Hashimoto's Thyroiditis on Ultrasound. J Clin Endocrinol Metab 2022; 107:953-963. [PMID: 34907442 PMCID: PMC8947219 DOI: 10.1210/clinem/dgab870] [Citation(s) in RCA: 18] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/27/2021] [Indexed: 02/05/2023]
Abstract
PURPOSE This study investigates the efficiency of deep learning models in the automated diagnosis of Hashimoto's thyroiditis (HT) using real-world ultrasound data from ultrasound examinations by computer-assisted diagnosis (CAD) with artificial intelligence. METHODS We retrospectively collected ultrasound images from patients with and without HT from 2 hospitals in China between September 2008 and February 2018. Images were divided into a training set (80%) and a validation set (20%). We ensembled 9 convolutional neural networks (CNNs) as the final model (CAD-HT) for HT classification. The model's diagnostic performance was validated and compared to 2 hospital validation sets. We also compared the accuracy of CAD-HT against seniors/junior radiologists. Subgroup analysis of CAD-HT performance for different thyroid hormone levels (hyperthyroidism, hypothyroidism, and euthyroidism) was also evaluated. RESULTS 39 280 ultrasound images from 21 118 patients were included in this study. The accuracy, sensitivity, and specificity of the HT-CAD model were 0.892, 0.890, and 0.895, respectively. HT-CAD performance between 2 hospitals was not significantly different. The HT-CAD model achieved a higher performance (P < 0.001) when compared to senior radiologists, with a nearly 9% accuracy improvement. HT-CAD had almost similar accuracy (range 0.87-0.894) for the 3 subgroups based on thyroid hormone level. CONCLUSION The HT-CAD strategy based on CNN significantly improved the radiologists' diagnostic accuracy of HT. Our model demonstrates good performance and robustness in different hospitals and for different thyroid hormone levels.
Collapse
Affiliation(s)
- Wanjun Zhao
- Department of Thyroid Surgery, West China Hospital, Sichuan University, Chengdu, China
| | - Qingbo Kang
- West China Biomedical Big Data Center, West China Hospital, Sichuan University, Chengdu, China
| | - Feiyan Qian
- Department of Rehabilitation, Shaoxing Central Hospital, Shaoxing, China
| | - Kang Li
- West China Biomedical Big Data Center, West China Hospital, Sichuan University, Chengdu, China
| | - Jingqiang Zhu
- Department of Thyroid Surgery, West China Hospital, Sichuan University, Chengdu, China
- Correspondence: Jingqiang Zhu, MD, Department of Thyroid Surgery, West China Hospital, Sichuan University, Chengdu 610041, China. ; or Buyun Ma, MD, Department of Ultrasonography, West China Hospital of Sichuan University, Chengdu 610041, China.
| | - Buyun Ma
- Department of Ultrasonography, West China Hospital, Sichuan University, Chengdu, China
| |
Collapse
|
24
|
Kunar MA. The optimal use of computer aided detection to find low prevalence cancers. Cogn Res Princ Implic 2022; 7:13. [PMID: 35122173 PMCID: PMC8816998 DOI: 10.1186/s41235-022-00361-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2021] [Accepted: 01/13/2022] [Indexed: 11/10/2022] Open
Abstract
People miss a high proportion of targets that only appear rarely. This low prevalence (LP) effect has implications for applied search tasks such as the clinical reading of mammograms. Computer aided detection (CAD) has been used to help radiologists search mammograms by highlighting areas likely to contain a cancer. Previous research has found a benefit in search when CAD cues were correct but a cost to search when CAD cues were incorrect. The current research investigated whether there is an optimal way to present CAD to ensure low error rates when CAD is both correct and incorrect. Experiment 1 compared an automatic condition, where CAD appeared simultaneously with the display to an interactive condition, where participants could choose to use CAD. Experiment 2 compared the automatic condition to a confirm condition, where participants searched the display first before being shown the CAD cues. The results showed that miss errors were reduced overall in the confirm condition, with no cost to false alarms. Furthermore, having CAD be interactive, resulted in a low uptake where it was only used in 34% of trials. The results showed that the presentation mode of CAD can affect decision-making in LP search.
Collapse
Affiliation(s)
- Melina A Kunar
- Department of Psychology, The University of Warwick, Coventry, CV4 7AL, UK.
| |
Collapse
|
25
|
Hickman SE, Woitek R, Le EPV, Im YR, Mouritsen Luxhøj C, Aviles-Rivero AI, Baxter GC, MacKay JW, Gilbert FJ. Machine Learning for Workflow Applications in Screening Mammography: Systematic Review and Meta-Analysis. Radiology 2022; 302:88-104. [PMID: 34665034 PMCID: PMC8717814 DOI: 10.1148/radiol.2021210391] [Citation(s) in RCA: 45] [Impact Index Per Article: 22.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2021] [Revised: 07/14/2021] [Accepted: 08/05/2021] [Indexed: 01/03/2023]
Abstract
Background Advances in computer processing and improvements in data availability have led to the development of machine learning (ML) techniques for mammographic imaging. Purpose To evaluate the reported performance of stand-alone ML applications for screening mammography workflow. Materials and Methods Ovid Embase, Ovid Medline, Cochrane Central Register of Controlled Trials, Scopus, and Web of Science literature databases were searched for relevant studies published from January 2012 to September 2020. The study was registered with the PROSPERO International Prospective Register of Systematic Reviews (protocol no. CRD42019156016). Stand-alone technology was defined as a ML algorithm that can be used independently of a human reader. Studies were quality assessed using the Quality Assessment of Diagnostic Accuracy Studies 2 and the Prediction Model Risk of Bias Assessment Tool, and reporting was evaluated using the Checklist for Artificial Intelligence in Medical Imaging. A primary meta-analysis included the top-performing algorithm and corresponding reader performance from which pooled summary estimates for the area under the receiver operating characteristic curve (AUC) were calculated using a bivariate model. Results Fourteen articles were included, which detailed 15 studies for stand-alone detection (n = 8) and triage (n = 7). Triage studies reported that 17%-91% of normal mammograms identified could be read by adapted screening, while "missing" an estimated 0%-7% of cancers. In total, an estimated 185 252 cases from three countries with more than 39 readers were included in the primary meta-analysis. The pooled sensitivity, specificity, and AUC was 75.4% (95% CI: 65.6, 83.2; P = .11), 90.6% (95% CI: 82.9, 95.0; P = .40), and 0.89 (95% CI: 0.84, 0.98), respectively, for algorithms, and 73.0% (95% CI: 60.7, 82.6), 88.6% (95% CI: 72.4, 95.8), and 0.85 (95% CI: 0.78, 0.97), respectively, for readers. Conclusion Machine learning (ML) algorithms that demonstrate a stand-alone application in mammographic screening workflows achieve or even exceed human reader detection performance and improve efficiency. However, this evidence is from a small number of retrospective studies. Therefore, further rigorous independent external prospective testing of ML algorithms to assess performance at preassigned thresholds is required to support these claims. ©RSNA, 2021 Online supplemental material is available for this article. See also the editorial by Whitman and Moseley in this issue.
Collapse
Affiliation(s)
- Sarah E. Hickman
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Ramona Woitek
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Elizabeth Phuong Vi Le
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Yu Ri Im
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Carina Mouritsen Luxhøj
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Angelica I. Aviles-Rivero
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Gabrielle C. Baxter
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - James W. MacKay
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Fiona J. Gilbert
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| |
Collapse
|
26
|
Artificial intelligence for the real world of breast screening. Eur J Radiol 2021; 144:109661. [PMID: 34598013 DOI: 10.1016/j.ejrad.2021.109661] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2021] [Revised: 03/08/2021] [Accepted: 03/15/2021] [Indexed: 11/21/2022]
Abstract
Breast cancer screening with mammography reduces mortality in the women who attend by detecting high risk cancer early. It is far from perfect with variations in both sensitivity for the detection of cancer and very wide variations in specificity, leading to unnecessary recalls and biopsies. Over the last 12 months several papers have reported on AI algorithms that perform as well as human readers on large well curated population data sets. The nature of the test sets, the way the gold standard has been calculated, the definition of a positive call, and the statistics used all influence the results. Historically retrospective studies have not predicted the real-life performance of radiologist plus machine. So, it is important to perform prospective studies before introducing Artificial intelligence into real world breast screening.
Collapse
|
27
|
Eng DK, Khandwala NB, Long J, Fefferman NR, Lala SV, Strubel NA, Milla SS, Filice RW, Sharp SE, Towbin AJ, Francavilla ML, Kaplan SL, Ecklund K, Prabhu SP, Dillon BJ, Everist BM, Anton CG, Bittman ME, Dennis R, Larson DB, Seekins JM, Silva CT, Zandieh AR, Langlotz CP, Lungren MP, Halabi SS. Artificial Intelligence Algorithm Improves Radiologist Performance in Skeletal Age Assessment: A Prospective Multicenter Randomized Controlled Trial. Radiology 2021; 301:692-699. [PMID: 34581608 DOI: 10.1148/radiol.2021204021] [Citation(s) in RCA: 43] [Impact Index Per Article: 14.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
Background Previous studies suggest that use of artificial intelligence (AI) algorithms as diagnostic aids may improve the quality of skeletal age assessment, though these studies lack evidence from clinical practice. Purpose To compare the accuracy and interpretation time of skeletal age assessment on hand radiograph examinations with and without the use of an AI algorithm as a diagnostic aid. Materials and Methods In this prospective randomized controlled trial, the accuracy of skeletal age assessment on hand radiograph examinations was performed with (n = 792) and without (n = 739) the AI algorithm as a diagnostic aid. For examinations with the AI algorithm, the radiologist was shown the AI interpretation as part of their routine clinical work and was permitted to accept or modify it. Hand radiographs were interpreted by 93 radiologists from six centers. The primary efficacy outcome was the mean absolute difference between the skeletal age dictated into the radiologists' signed report and the average interpretation of a panel of four radiologists not using a diagnostic aid. The secondary outcome was the interpretation time. A linear mixed-effects regression model with random center- and radiologist-level effects was used to compare the two experimental groups. Results Overall mean absolute difference was lower when radiologists used the AI algorithm compared with when they did not (5.36 months vs 5.95 months; P = .04). The proportions at which the absolute difference exceeded 12 months (9.3% vs 13.0%, P = .02) and 24 months (0.5% vs 1.8%, P = .02) were lower with the AI algorithm than without it. Median radiologist interpretation time was lower with the AI algorithm than without it (102 seconds vs 142 seconds, P = .001). Conclusion Use of an artificial intelligence algorithm improved skeletal age assessment accuracy and reduced interpretation times for radiologists, although differences were observed between centers. Clinical trial registration no. NCT03530098 © RSNA, 2021 Online supplemental material is available for this article. See also the editorial by Rubin in this issue.
Collapse
Affiliation(s)
- David K Eng
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Nishith B Khandwala
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Jin Long
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Nancy R Fefferman
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Shailee V Lala
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Naomi A Strubel
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Sarah S Milla
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Ross W Filice
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Susan E Sharp
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Alexander J Towbin
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Michael L Francavilla
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Summer L Kaplan
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Kirsten Ecklund
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Sanjay P Prabhu
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Brian J Dillon
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Brian M Everist
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Christopher G Anton
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Mark E Bittman
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Rebecca Dennis
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - David B Larson
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Jayne M Seekins
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Cicero T Silva
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Arash R Zandieh
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Curtis P Langlotz
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Matthew P Lungren
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| | - Safwan S Halabi
- From the Department of Computer Science, Stanford University, 300 N Pasteur Dr, Stanford, CA 94305 (D.K.E., N.B.K.); Departments of Pediatrics (J.L.) and Radiology (D.B.L., J.M.S., C.P.L., M.P.L., S.S.H.), Stanford University School of Medicine, Stanford, Calif; Department of Radiology, New York University School of Medicine, New York, NY (N.R.F., S.V.L., N.A.S., M.E.B.); Department of Radiology, Emory School of Medicine and Children's Healthcare of Atlanta, Atlanta, Ga (S.S.M.); Department of Radiology, MedStar Health and Georgetown University School of Medicine, Washington, DC (R.W.F., A.R.Z.); Department of Radiology, Cincinnati Children's Hospital Medical Center, Cincinnati, Ohio (S.E.S., A.J.T., C.G.A.); Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, Pa (M.L.F., S.L.K., R.D.); Department of Radiology, Harvard Medical School and Boston Children's Hospital, Boston, Mass (K.E., S.P.P.); Department of Radiology, Yale School of Medicine, New Haven, Conn (B.J.D., C.T.S.); and Department of Radiology, Kansas University School of Medicine, Kansas City, Kan (B.M.E.)
| |
Collapse
|
28
|
Mori Y, Bretthauer M, Kalager M. Hopes and Hypes for Artificial Intelligence in Colorectal Cancer Screening. Gastroenterology 2021; 161:774-777. [PMID: 33989659 DOI: 10.1053/j.gastro.2021.04.078] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/14/2021] [Revised: 04/20/2021] [Accepted: 04/26/2021] [Indexed: 12/13/2022]
Affiliation(s)
- Yuichi Mori
- Clinical Effectiveness Research Group, University of Oslo, Oslo, Norway; Digestive Disease Center, Showa University Northern Yokohama Hospital, Yokohama, Japan
| | - Michael Bretthauer
- Clinical Effectiveness Research Group, University of Oslo; Department of Transplantation Medicine, Oslo University Hospital, Oslo, Norway
| | - Mette Kalager
- Clinical Effectiveness Research Group, University of Oslo; Department of Transplantation Medicine, Oslo University Hospital, Oslo, Norway
| |
Collapse
|
29
|
Hickman SE, Baxter GC, Gilbert FJ. Adoption of artificial intelligence in breast imaging: evaluation, ethical constraints and limitations. Br J Cancer 2021; 125:15-22. [PMID: 33772149 PMCID: PMC8257639 DOI: 10.1038/s41416-021-01333-w] [Citation(s) in RCA: 47] [Impact Index Per Article: 15.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2020] [Revised: 02/15/2021] [Accepted: 02/24/2021] [Indexed: 02/07/2023] Open
Abstract
Retrospective studies have shown artificial intelligence (AI) algorithms can match as well as enhance radiologist's performance in breast screening. These tools can facilitate tasks not feasible by humans such as the automatic triage of patients and prediction of treatment outcomes. Breast imaging faces growing pressure with the exponential growth in imaging requests and a predicted reduced workforce to provide reports. Solutions to alleviate these pressures are being sought with an increasing interest in the adoption of AI to improve workflow efficiency as well as patient outcomes. Vast quantities of data are needed to test and monitor AI algorithms before and after their incorporation into healthcare systems. Availability of data is currently limited, although strategies are being devised to harness the data that already exists within healthcare institutions. Challenges that underpin the realisation of AI into everyday breast imaging cannot be underestimated and the provision of guidance from national agencies to tackle these challenges, taking into account views from a societal, industrial and healthcare prospective is essential. This review provides background on the evaluation and use of AI in breast imaging in addition to exploring key ethical, technical, legal and regulatory challenges that have been identified so far.
Collapse
Affiliation(s)
- Sarah E Hickman
- Department of Radiology, University of Cambridge School of Clinical Medicine, Cambridge, UK
| | - Gabrielle C Baxter
- Department of Radiology, University of Cambridge School of Clinical Medicine, Cambridge, UK
| | - Fiona J Gilbert
- Department of Radiology, University of Cambridge School of Clinical Medicine, Cambridge, UK.
- Department of Radiology, Addenbrookes Hospital, Cambridge University Hospitals NHS Foundation Trust, Cambridge, UK.
| |
Collapse
|
30
|
Murphy MP, Brown NM. CORR Synthesis: When Should the Orthopaedic Surgeon Use Artificial Intelligence, Machine Learning, and Deep Learning? Clin Orthop Relat Res 2021; 479:1497-1505. [PMID: 33595930 PMCID: PMC8208440 DOI: 10.1097/corr.0000000000001679] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/10/2020] [Accepted: 01/22/2021] [Indexed: 01/31/2023]
Affiliation(s)
- Michael P Murphy
- Department of Orthopaedic Surgery and Rehabilitation, Loyola University Medical Center, Maywood, IL, USA
| | | |
Collapse
|
31
|
Yoon JH, Kim EK. Deep Learning-Based Artificial Intelligence for Mammography. Korean J Radiol 2021; 22:1225-1239. [PMID: 33987993 PMCID: PMC8316774 DOI: 10.3348/kjr.2020.1210] [Citation(s) in RCA: 23] [Impact Index Per Article: 7.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2020] [Revised: 01/11/2021] [Accepted: 01/17/2021] [Indexed: 12/27/2022] Open
Abstract
During the past decade, researchers have investigated the use of computer-aided mammography interpretation. With the application of deep learning technology, artificial intelligence (AI)-based algorithms for mammography have shown promising results in the quantitative assessment of parenchymal density, detection and diagnosis of breast cancer, and prediction of breast cancer risk, enabling more precise patient management. AI-based algorithms may also enhance the efficiency of the interpretation workflow by reducing both the workload and interpretation time. However, more in-depth investigation is required to conclusively prove the effectiveness of AI-based algorithms. This review article discusses how AI algorithms can be applied to mammography interpretation as well as the current challenges in its implementation in real-world practice.
Collapse
Affiliation(s)
- Jung Hyun Yoon
- Department of Radiology, Severance Hospital, Research Institute of Radiological Science, Seoul, Korea
| | - Eun Kyung Kim
- Department of Radiology, Yongin Severance Hospital, Yonsei University, College of Medicine, Yongin, Korea.
| |
Collapse
|
32
|
Lorkowski J, Grzegorowska O, Pokorski M. Artificial Intelligence in the Healthcare System: An Overview. ADVANCES IN EXPERIMENTAL MEDICINE AND BIOLOGY 2021; 1335:1-10. [PMID: 33768498 DOI: 10.1007/5584_2021_620] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Subscribe] [Scholar Register] [Indexed: 01/05/2023]
Abstract
This chapter aims to present insights into the influence of artificial intelligence (AI) on medicine, public health, and the economy. PubMed and Google Scholar databases were used for the identification and collection of articles with search commands of "artificial intelligence" AND "public health" and "artificial intelligence" AND "medicine". A total of 273 articles specifically handling the issue of artificial intelligence, dating ten years back, in three major medical journals: Science, The Lancet, and The New England Journal of Medicine, were analyzed. Computational power gets stronger by the day, giving us new solutions and possibilities. Current medicine problems like personalized medicine, storage of data, and documentation overload will likely be replaced by AI shortly. The application of AI may also bring substantial benefits to other areas of medicine like the diagnostic and therapeutic processes. The development and spread of AI are inescapable as it lowers healthcare and administrative costs, improves medical efficiency, and predicts and prevents major disease complications. The use of AI in medicine seems destined to carry the day.
Collapse
Affiliation(s)
- Jacek Lorkowski
- Department of Orthopedics, Traumatology and Sports Medicine, Central Clinical Hospital of the Ministry of Internal Affairs and Administration, Warsaw, Poland. .,Faculty of Health Sciences, Medical University of Mazovia, Warsaw, Poland.
| | - Oliwia Grzegorowska
- Department of Cardiology, Independent Public Regional Hospital, Szczecin, Poland
| | - Mieczysław Pokorski
- Faculty of Health Sciences, The Jan Długosz University in Częstochowa, Częstochowa, Poland.,Institute of Health Sciences, Opole University, Opole, Poland
| |
Collapse
|
33
|
Chan HP, Hadjiiski LM, Samala RK. Computer-aided diagnosis in the era of deep learning. Med Phys 2021; 47:e218-e227. [PMID: 32418340 DOI: 10.1002/mp.13764] [Citation(s) in RCA: 99] [Impact Index Per Article: 33.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2019] [Revised: 05/13/2019] [Accepted: 05/13/2019] [Indexed: 12/15/2022] Open
Abstract
Computer-aided diagnosis (CAD) has been a major field of research for the past few decades. CAD uses machine learning methods to analyze imaging and/or nonimaging patient data and makes assessment of the patient's condition, which can then be used to assist clinicians in their decision-making process. The recent success of the deep learning technology in machine learning spurs new research and development efforts to improve CAD performance and to develop CAD for many other complex clinical tasks. In this paper, we discuss the potential and challenges in developing CAD tools using deep learning technology or artificial intelligence (AI) in general, the pitfalls and lessons learned from CAD in screening mammography and considerations needed for future implementation of CAD or AI in clinical use. It is hoped that the past experiences and the deep learning technology will lead to successful advancement and lasting growth in this new era of CAD, thereby enabling CAD to deliver intelligent aids to improve health care.
Collapse
Affiliation(s)
- Heang-Ping Chan
- Department of Radiology, University of Michigan, Ann Arbor, MI, 48109-5842, USA
| | - Lubomir M Hadjiiski
- Department of Radiology, University of Michigan, Ann Arbor, MI, 48109-5842, USA
| | - Ravi K Samala
- Department of Radiology, University of Michigan, Ann Arbor, MI, 48109-5842, USA
| |
Collapse
|
34
|
AI applications in robotics, diagnostic image analysis and precision medicine: Current limitations, future trends, guidelines on CAD systems for medicine. INFORMATICS IN MEDICINE UNLOCKED 2021. [DOI: 10.1016/j.imu.2021.100596] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/22/2022] Open
|
35
|
Džoić Dominković M, Ivanac G, Radović N, Čavka M. WHAT CAN WE ACTUALLY SEE USING COMPUTER AIDED DETECTION IN MAMMOGRAPHY? Acta Clin Croat 2020; 59:576-581. [PMID: 34285427 PMCID: PMC8253062 DOI: 10.20471/acc.2020.59.04.02] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2018] [Accepted: 04/12/2018] [Indexed: 11/24/2022] Open
Abstract
The main goal of this study was to compare the results of computer aided detection (CAD) analysis in screening mammography with the results independently obtained by two radiologists for the same samples and to determine the sensitivity and specificity of CAD for breast lesions. A total of 436 mammograms were analyzed with CAD. For each screening mammogram, the changes in breast tissue recognized by CAD were compared to the interpretations of two radiologists. The sensitivity and specificity of CAD for breast lesions were calculated using contingency table. The sensitivity of CAD for all lesions was 54% and specificity 16%. CAD sensitivity for suspicious lesions only was 86%. CAD sensitivity for microcalcifications was 100% and specificity 45%. CAD mainly ‘mistook’ glandular parenchyma, connective tissue and blood vessels for breast lesions, and blood vessel calcifications and axillary folds for microcalcifications. In this study, we confirmed CAD as an excellent tool for recognizing microcalcifications with 100% sensitivity. However, it should not be used as a stand-alone tool in breast screening mammography due to the high rate of false-positive results.
Collapse
Affiliation(s)
- Martina Džoić Dominković
- 1Department of Radiology, Orašje General Hospital, Orašje, Bosnia and Herzegovina; 2Department of Diagnostic and Interventional Radiology, Dubrava University Hospital, Zagreb, Croatia; 3University of Zagreb, School of Medicine, Zagreb, Croatia
| | - Gordana Ivanac
- 1Department of Radiology, Orašje General Hospital, Orašje, Bosnia and Herzegovina; 2Department of Diagnostic and Interventional Radiology, Dubrava University Hospital, Zagreb, Croatia; 3University of Zagreb, School of Medicine, Zagreb, Croatia
| | - Niko Radović
- 1Department of Radiology, Orašje General Hospital, Orašje, Bosnia and Herzegovina; 2Department of Diagnostic and Interventional Radiology, Dubrava University Hospital, Zagreb, Croatia; 3University of Zagreb, School of Medicine, Zagreb, Croatia
| | - Mislav Čavka
- 1Department of Radiology, Orašje General Hospital, Orašje, Bosnia and Herzegovina; 2Department of Diagnostic and Interventional Radiology, Dubrava University Hospital, Zagreb, Croatia; 3University of Zagreb, School of Medicine, Zagreb, Croatia
| |
Collapse
|
36
|
Does Artificial Intelligence Outperform Natural Intelligence in Interpreting Musculoskeletal Radiological Studies? A Systematic Review. Clin Orthop Relat Res 2020; 478:2751-2764. [PMID: 32740477 PMCID: PMC7899420 DOI: 10.1097/corr.0000000000001360] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 01/31/2023]
Abstract
BACKGROUND Machine learning (ML) is a subdomain of artificial intelligence that enables computers to abstract patterns from data without explicit programming. A myriad of impactful ML applications already exists in orthopaedics ranging from predicting infections after surgery to diagnostic imaging. However, no systematic reviews that we know of have compared, in particular, the performance of ML models with that of clinicians in musculoskeletal imaging to provide an up-to-date summary regarding the extent of applying ML to imaging diagnoses. By doing so, this review delves into where current ML developments stand in aiding orthopaedists in assessing musculoskeletal images. QUESTIONS/PURPOSES This systematic review aimed (1) to compare performance of ML models versus clinicians in detecting, differentiating, or classifying orthopaedic abnormalities on imaging by (A) accuracy, sensitivity, and specificity, (B) input features (for example, plain radiographs, MRI scans, ultrasound), (C) clinician specialties, and (2) to compare the performance of clinician-aided versus unaided ML models. METHODS A systematic review was performed in PubMed, Embase, and the Cochrane Library for studies published up to October 1, 2019, using synonyms for machine learning and all potential orthopaedic specialties. We included all studies that compared ML models head-to-head against clinicians in the binary detection of abnormalities in musculoskeletal images. After screening 6531 studies, we ultimately included 12 studies. We conducted quality assessment using the Methodological Index for Non-randomized Studies (MINORS) checklist. All 12 studies were of comparable quality, and they all clearly included six of the eight critical appraisal items (study aim, input feature, ground truth, ML versus human comparison, performance metric, and ML model description). This justified summarizing the findings in a quantitative form by calculating the median absolute improvement of the ML models compared with clinicians for the following metrics of performance: accuracy, sensitivity, and specificity. RESULTS ML models provided, in aggregate, only very slight improvements in diagnostic accuracy and sensitivity compared with clinicians working alone and were on par in specificity (3% (interquartile range [IQR] -2.0% to 7.5%), 0.06% (IQR -0.03 to 0.14), and 0.00 (IQR -0.048 to 0.048), respectively). Inputs used by the ML models were plain radiographs (n = 8), MRI scans (n = 3), and ultrasound examinations (n = 1). Overall, ML models outperformed clinicians more when interpreting plain radiographs than when interpreting MRIs (17 of 34 and 3 of 16 performance comparisons, respectively). Orthopaedists and radiologists performed similarly to ML models, while ML models mostly outperformed other clinicians (outperformance in 7 of 19, 7 of 23, and 6 of 10 performance comparisons, respectively). Two studies evaluated the performance of clinicians aided and unaided by ML models; both demonstrated considerable improvements in ML-aided clinician performance by reporting a 47% decrease of misinterpretation rate (95% confidence interval [CI] 37 to 54; p < 0.001) and a mean increase in specificity of 0.048 (95% CI 0.029 to 0.068; p < 0.001) in detecting abnormalities on musculoskeletal images. CONCLUSIONS At present, ML models have comparable performance to clinicians in assessing musculoskeletal images. ML models may enhance the performance of clinicians as a technical supplement rather than as a replacement for clinical intelligence. Future ML-related studies should emphasize how ML models can complement clinicians, instead of determining the overall superiority of one versus the other. This can be accomplished by improving transparent reporting, diminishing bias, determining the feasibility of implantation in the clinical setting, and appropriately tempering conclusions. LEVEL OF EVIDENCE Level III, diagnostic study.
Collapse
|
37
|
Pacilè S, Lopez J, Chone P, Bertinotti T, Grouin JM, Fillard P. Improving Breast Cancer Detection Accuracy of Mammography with the Concurrent Use of an Artificial Intelligence Tool. Radiol Artif Intell 2020; 2:e190208. [PMID: 33937844 DOI: 10.1148/ryai.2020190208] [Citation(s) in RCA: 55] [Impact Index Per Article: 13.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2019] [Revised: 06/27/2020] [Accepted: 07/07/2020] [Indexed: 02/06/2023]
Abstract
Purpose To evaluate the benefits of an artificial intelligence (AI)-based tool for two-dimensional mammography in the breast cancer detection process. Materials and Methods In this multireader, multicase retrospective study, 14 radiologists assessed a dataset of 240 digital mammography images, acquired between 2013 and 2016, using a counterbalance design in which half of the dataset was read without AI and the other half with the help of AI during a first session and vice versa during a second session, which was separated from the first by a washout period. Area under the receiver operating characteristic curve (AUC), sensitivity, specificity, and reading time were assessed as endpoints. Results The average AUC across readers was 0.769 (95% CI: 0.724, 0.814) without AI and 0.797 (95% CI: 0.754, 0.840) with AI. The average difference in AUC was 0.028 (95% CI: 0.002, 0.055, P = .035). Average sensitivity was increased by 0.033 when using AI support (P = .021). Reading time changed dependently to the AI-tool score. For low likelihood of malignancy (< 2.5%), the time was about the same in the first reading session and slightly decreased in the second reading session. For higher likelihood of malignancy, the reading time was on average increased with the use of AI. Conclusion This clinical investigation demonstrated that the concurrent use of this AI tool improved the diagnostic performance of radiologists in the detection of breast cancer without prolonging their workflow.Supplemental material is available for this article.© RSNA, 2020.
Collapse
Affiliation(s)
- Serena Pacilè
- Therapixel SA, 39 Rue Claude Daunesse, 06560 Valbonne, France (S.P., P.C., T.B., P.F.); Radiology & Imaging Services, Hoag Memorial Hospital Presbyterian, Newport Beach, Calif (J.L.); and Department of Statistics, University of Rouen, Rouen, France (J.M.G.)
| | - January Lopez
- Therapixel SA, 39 Rue Claude Daunesse, 06560 Valbonne, France (S.P., P.C., T.B., P.F.); Radiology & Imaging Services, Hoag Memorial Hospital Presbyterian, Newport Beach, Calif (J.L.); and Department of Statistics, University of Rouen, Rouen, France (J.M.G.)
| | - Pauline Chone
- Therapixel SA, 39 Rue Claude Daunesse, 06560 Valbonne, France (S.P., P.C., T.B., P.F.); Radiology & Imaging Services, Hoag Memorial Hospital Presbyterian, Newport Beach, Calif (J.L.); and Department of Statistics, University of Rouen, Rouen, France (J.M.G.)
| | - Thomas Bertinotti
- Therapixel SA, 39 Rue Claude Daunesse, 06560 Valbonne, France (S.P., P.C., T.B., P.F.); Radiology & Imaging Services, Hoag Memorial Hospital Presbyterian, Newport Beach, Calif (J.L.); and Department of Statistics, University of Rouen, Rouen, France (J.M.G.)
| | - Jean Marie Grouin
- Therapixel SA, 39 Rue Claude Daunesse, 06560 Valbonne, France (S.P., P.C., T.B., P.F.); Radiology & Imaging Services, Hoag Memorial Hospital Presbyterian, Newport Beach, Calif (J.L.); and Department of Statistics, University of Rouen, Rouen, France (J.M.G.)
| | - Pierre Fillard
- Therapixel SA, 39 Rue Claude Daunesse, 06560 Valbonne, France (S.P., P.C., T.B., P.F.); Radiology & Imaging Services, Hoag Memorial Hospital Presbyterian, Newport Beach, Calif (J.L.); and Department of Statistics, University of Rouen, Rouen, France (J.M.G.)
| |
Collapse
|
38
|
Goto Y, Tanaka R, Furuya Y, Maezato M, Akita F, Shiraishi J. [Investigation of Clinical Utility of Radiological Technologist's Reading Report as a Second Opinion for Medical Doctor Reading of Digital Mammogram]. Nihon Hoshasen Gijutsu Gakkai Zasshi 2020; 76:997-1008. [PMID: 33087659 DOI: 10.6009/jjrt.2020_jsrt_76.10.997] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
PURPOSE We investigated the clinical utility of a radiological technologist's (RT)'s reports (RRs) as a second opinion by the free-response receiver operating characteristic (FROC) observer study that compared the performance of medical doctors' (MDs') reading of digital mammogram with and without consulting the RR. METHOD One hundred women (39 malignant, 61 benign or normal) who underwent diagnostic mammography were selected from among 1674 routine clinical images classified by the degree of difficulty and categories for inclusion in the FROC study. The first FROC study performed by three RTs (RT 1-3) was conducted to collect the data for RR utilized in the second FROC study. The second FROC study was performed by five MDs, and the statistical significance of MDs' performances with and without reference to the RR was investigated by figure of merit (FOM). RESULT The FOM values of three RTs obtained in the first FROC study were 0.529, 0.576, and 0.539, respectively. In the second FROC study, RT 2 had the highest FOM, RT 1 the lowest false positives/case, and RT 3 the highest sensitivity. The average FOM values in the second FROC study for the five MDs with/without reference to the RR were as follows: RT 2's RR was 0.534/0.588 (p=0.003), RT 1's RR was 0.500/0.545 (p=0.099), and RT 3's RR was 0.569/0.592 (p=0.324). CONCLUSION We concluded that the MDs' performance of reading mammogram was statistically improved by consulting the RR when the RT's reading skill was high.
Collapse
Affiliation(s)
- Yuka Goto
- Breast and Imaging Center of St. Marianna University School of Medicine
| | - Rie Tanaka
- School of Health Sciences, College of Medical, Pharmaceutical and Health Sciences, Kanazawa University
| | - Yuko Furuya
- Breast and Imaging Center of St. Marianna University School of Medicine
| | - Miwako Maezato
- Imaging Center, St. Marianna University School of Medicine
| | | | | |
Collapse
|
39
|
Al'Aref SJ, Anchouche K, Singh G, Slomka PJ, Kolli KK, Kumar A, Pandey M, Maliakal G, van Rosendael AR, Beecy AN, Berman DS, Leipsic J, Nieman K, Andreini D, Pontone G, Schoepf UJ, Shaw LJ, Chang HJ, Narula J, Bax JJ, Guan Y, Min JK. Clinical applications of machine learning in cardiovascular disease and its relevance to cardiac imaging. Eur Heart J 2020; 40:1975-1986. [PMID: 30060039 DOI: 10.1093/eurheartj/ehy404] [Citation(s) in RCA: 233] [Impact Index Per Article: 58.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/21/2018] [Revised: 05/29/2018] [Accepted: 07/06/2018] [Indexed: 12/19/2022] Open
Abstract
Artificial intelligence (AI) has transformed key aspects of human life. Machine learning (ML), which is a subset of AI wherein machines autonomously acquire information by extracting patterns from large databases, has been increasingly used within the medical community, and specifically within the domain of cardiovascular diseases. In this review, we present a brief overview of ML methodologies that are used for the construction of inferential and predictive data-driven models. We highlight several domains of ML application such as echocardiography, electrocardiography, and recently developed non-invasive imaging modalities such as coronary artery calcium scoring and coronary computed tomography angiography. We conclude by reviewing the limitations associated with contemporary application of ML algorithms within the cardiovascular disease field.
Collapse
Affiliation(s)
- Subhi J Al'Aref
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Khalil Anchouche
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Gurpreet Singh
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Piotr J Slomka
- Departments of Imaging and Medicine and Biomedical Sciences, Cedars-Sinai Medical Center, Los Angeles, CA, USA
| | - Kranthi K Kolli
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Amit Kumar
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Mohit Pandey
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Gabriel Maliakal
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Alexander R van Rosendael
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Ashley N Beecy
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Daniel S Berman
- Departments of Imaging and Medicine and Biomedical Sciences, Cedars-Sinai Medical Center, Los Angeles, CA, USA
| | - Jonathan Leipsic
- Departments of Medicine and Radiology, University of British Columbia, Vancouver, BC, Canada
| | - Koen Nieman
- Departments of Cardiology and Radiology, Stanford University School of Medicine and Cardiovascular Institute, Stanford, CA, USA
| | | | | | - U Joseph Schoepf
- Division of Cardiovascular Imaging, Department of Radiology and Radiological Science and Division of Cardiology, Department of Medicine, Medical University of South Carolina, Charleston, SC, USA
| | - Leslee J Shaw
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| | - Hyuk-Jae Chang
- Division of Cardiology, Severance Cardiovascular Hospital and Severance Biomedical Science Institute, Yonsei University College of Medicine, Yonsei University Health System, Seoul, South Korea
| | - Jagat Narula
- Zena and Michael A. Wiener Cardiovascular Institute, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| | - Jeroen J Bax
- Department of Cardiology, Heart Lung Center, Leiden University Medical Center, Leiden, The Netherlands
| | - Yuanfang Guan
- Department of Computational Medicine and Bioinformatics, University of Michigan, Ann Arbor, MI, USA
| | - James K Min
- Department of Radiology, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY, USA
| |
Collapse
|
40
|
Hwang EJ, Park S, Jin KN, Kim JI, Choi SY, Lee JH, Goo JM, Aum J, Yim JJ, Park CM. Development and Validation of a Deep Learning-based Automatic Detection Algorithm for Active Pulmonary Tuberculosis on Chest Radiographs. Clin Infect Dis 2020; 69:739-747. [PMID: 30418527 PMCID: PMC6695514 DOI: 10.1093/cid/ciy967] [Citation(s) in RCA: 108] [Impact Index Per Article: 27.0] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2018] [Accepted: 11/08/2018] [Indexed: 12/25/2022] Open
Abstract
Background Detection of active pulmonary tuberculosis on chest radiographs (CRs) is critical for the diagnosis and screening of tuberculosis. An automated system may help streamline the tuberculosis screening process and improve diagnostic performance. Methods We developed a deep learning–based automatic detection (DLAD) algorithm using 54c221 normal CRs and 6768 CRs with active pulmonary tuberculosis that were labeled and annotated by 13 board-certified radiologists. The performance of DLAD was validated using 6 external multicenter, multinational datasets. To compare the performances of DLAD with physicians, an observer performance test was conducted by 15 physicians including nonradiology physicians, board-certified radiologists, and thoracic radiologists. Image-wise classification and lesion-wise localization performances were measured using area under the receiver operating characteristic (ROC) curves and area under the alternative free-response ROC curves, respectively. Sensitivities and specificities of DLAD were calculated using 2 cutoffs (high sensitivity [98%] and high specificity [98%]) obtained through in-house validation. Results DLAD demonstrated classification performance of 0.977–1.000 and localization performance of 0.973–1.000. Sensitivities and specificities for classification were 94.3%–100% and 91.1%–100% using the high-sensitivity cutoff and 84.1%–99.0% and 99.1%–100% using the high-specificity cutoff. DLAD showed significantly higher performance in both classification (0.993 vs 0.746–0.971) and localization (0.993 vs 0.664–0.925) compared to all groups of physicians. Conclusions Our DLAD demonstrated excellent and consistent performance in the detection of active pulmonary tuberculosis on CR, outperforming physicians, including thoracic radiologists.
Collapse
Affiliation(s)
- Eui Jin Hwang
- Department of Radiology, Seoul National University College of Medicine, Seoul
| | - Sunggyun Park
- Lunit Inc, Seoul National University Boramae Medical Center, Seoul
| | - Kwang-Nam Jin
- Department of Radiology, Seoul National University Boramae Medical Center, Seoul
| | - Jung Im Kim
- Department of Radiology, Kyung Hee University Hospital at Gangdong, Seoul
| | - So Young Choi
- Department of Radiology, Eulji University Medical Center, Daejon
| | - Jong Hyuk Lee
- Department of Radiology, Seoul National University College of Medicine, Seoul
| | - Jin Mo Goo
- Department of Radiology, Seoul National University College of Medicine, Seoul
| | - Jaehong Aum
- Lunit Inc, Seoul National University Boramae Medical Center, Seoul
| | - Jae-Joon Yim
- Division of Pulmonary and Critical Care Medicine, Department of Internal Medicine, Seoul National University College of Medicine, Korea
| | - Chang Min Park
- Department of Radiology, Seoul National University College of Medicine, Seoul
| | | |
Collapse
|
41
|
Gale A, Chen Y. A review of the PERFORMS scheme in breast screening. Br J Radiol 2020; 93:20190908. [PMID: 32501766 PMCID: PMC7446009 DOI: 10.1259/bjr.20190908] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2019] [Revised: 05/22/2020] [Accepted: 05/29/2020] [Indexed: 11/05/2022] Open
Abstract
This review details the aetiology of the PERFORMS self-assessment scheme in breast screening, together with its subsequent development, current implementation and future function. The purpose of the scheme is examined and the importance of its continuing role in a changing screening service described, together with current evolution.
Collapse
Affiliation(s)
- Alastair Gale
- Department of Computer Science, Loughborough University, Loughborough, UK
| | - Yan Chen
- Division of Cancer and Stem Cells, University of Nottingham, Nottingham, UK
| |
Collapse
|
42
|
Al'Aref SJ, Singh G, van Rosendael AR, Kolli KK, Ma X, Maliakal G, Pandey M, Lee BC, Wang J, Xu Z, Zhang Y, Min JK, Wong SC, Minutello RM. Determinants of In-Hospital Mortality After Percutaneous Coronary Intervention: A Machine Learning Approach. J Am Heart Assoc 2020; 8:e011160. [PMID: 30834806 PMCID: PMC6474922 DOI: 10.1161/jaha.118.011160] [Citation(s) in RCA: 43] [Impact Index Per Article: 10.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 01/16/2023]
Abstract
Background The ability to accurately predict the occurrence of in‐hospital death after percutaneous coronary intervention is important for clinical decision‐making. We sought to utilize the New York Percutaneous Coronary Intervention Reporting System in order to elucidate the determinants of in‐hospital mortality in patients undergoing percutaneous coronary intervention across New York State. Methods and Results We examined 479 804 patients undergoing percutaneous coronary intervention between 2004 and 2012, utilizing traditional and advanced machine learning algorithms to determine the most significant predictors of in‐hospital mortality. The entire data were randomly split into a training (80%) and a testing set (20%). Tuned hyperparameters were used to generate a trained model while the performance of the model was independently evaluated on the testing set after plotting a receiver‐operator characteristic curve and using the output measure of the area under the curve (AUC) and the associated 95% CIs. Mean age was 65.2±11.9 years and 68.5% were women. There were 2549 in‐hospital deaths within the patient population. A boosted ensemble algorithm (AdaBoost) had optimal discrimination with AUC of 0.927 (95% CI 0.923–0.929) compared with AUC of 0.913 for XGBoost (95% CI 0.906–0.919, P=0.02), AUC of 0.892 for Random Forest (95% CI 0.889–0.896, P<0.01), and AUC of 0.908 for logistic regression (95% CI 0.907–0.910, P<0.01). The 2 most significant predictors were age and ejection fraction. Conclusions A big data approach that utilizes advanced machine learning algorithms identifies new associations among risk factors and provides high accuracy for the prediction of in‐hospital mortality in patients undergoing percutaneous coronary intervention. See Editorial by Garratt and Schneider
Collapse
Affiliation(s)
- Subhi J Al'Aref
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | - Gurpreet Singh
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | | | - Kranthi K Kolli
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | - Xiaoyue Ma
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | - Gabriel Maliakal
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | - Mohit Pandey
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | - Bejamin C Lee
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | - Jing Wang
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | - Zhuoran Xu
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | - Yiye Zhang
- 2 Division of Health Informatics Weill Cornell Graduate School of Medical Sciences New York NY
| | - James K Min
- 1 Dalio Institute of Cardiovascular Imaging New York-Presbyterian Hospital New York NY
| | - S Chiu Wong
- 3 Division of Cardiology Department of Medicine Weill Cornell Medicine New York NY
| | - Robert M Minutello
- 3 Division of Cardiology Department of Medicine Weill Cornell Medicine New York NY
| |
Collapse
|
43
|
Yamaguchi T, Inoue K, Tsunoda H, Uematsu T, Shinohara N, Mukai H. A deep learning-based automated diagnostic system for classifying mammographic lesions. Medicine (Baltimore) 2020; 99:e20977. [PMID: 32629712 PMCID: PMC7337553 DOI: 10.1097/md.0000000000020977] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/26/2022] Open
Abstract
BACKGROUND Screening mammography has led to reduced breast cancer-specific mortality and is recommended worldwide. However, the resultant doctors' workload of reading mammographic scans needs to be addressed. Although computer-aided detection (CAD) systems have been developed to support readers, the findings are conflicting regarding whether traditional CAD systems improve reading performance. Rapid progress in the artificial intelligence (AI) field has led to the advent of newer CAD systems using deep learning-based algorithms which have the potential to reach human performance levels. Those systems, however, have been developed using mammography images mainly from women in western countries. Because Asian women characteristically have higher-density breasts, it is uncertain whether those AI systems can apply to Japanese women. In this study, we will construct a deep learning-based CAD system trained using mammography images from a large number of Japanese women with high quality reading. METHODS We will collect digital mammography images taken for screening or diagnostic purposes at multiple institutions in Japan. A total of 15,000 images, consisting of 5000 images with breast cancer and 10,000 images with benign lesions, will be collected. At least 1000 images of normal breasts will also be collected for use as reference data. With these data, we will construct a deep learning-based AI system to detect breast cancer on mammograms. The primary endpoint will be the sensitivity and specificity of the AI system with the test image set. DISCUSSION When the ability of AI reading is shown to be on a par with that of human reading, images of normal breasts or benign lesions that do not have to be read by a human can be selected by AI beforehand. Our AI might work well in Asian women who have similar breast density, size, and shape to those of Japanese women. TRIAL REGISTRATION UMIN, trial number UMIN000039009. Registered 26 December 2019, https://www.umin.ac.jp/ctr/.
Collapse
Affiliation(s)
| | - Kenichi Inoue
- Breast Cancer Center, Shonan Memorial Hospital, Kanagawa
| | - Hiroko Tsunoda
- Department of Radiology, St. Luke's International Hospital, Tokyo
| | - Takayoshi Uematsu
- Division of Breast Imaging and Breast Interventional Radiology, Shizuoka Cancer Center Hospital, Shizuoka
| | - Norimitsu Shinohara
- Department of Radiological Technology, Faculty of Health Sciences, Gifu University of Medical Science, Gifu
| | - Hirofumi Mukai
- Division of Breast and Medical Oncology, National Cancer Center Hospital East, Chiba, Japan
| |
Collapse
|
44
|
Rodriguez-Ruiz A, Lång K, Gubern-Merida A, Broeders M, Gennaro G, Clauser P, Helbich TH, Chevalier M, Tan T, Mertelmeier T, Wallis MG, Andersson I, Zackrisson S, Mann RM, Sechopoulos I. Stand-Alone Artificial Intelligence for Breast Cancer Detection in Mammography: Comparison With 101 Radiologists. J Natl Cancer Inst 2020; 111:916-922. [PMID: 30834436 DOI: 10.1093/jnci/djy222] [Citation(s) in RCA: 294] [Impact Index Per Article: 73.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2018] [Revised: 10/06/2018] [Accepted: 11/29/2018] [Indexed: 12/11/2022] Open
Abstract
BACKGROUND Artificial intelligence (AI) systems performing at radiologist-like levels in the evaluation of digital mammography (DM) would improve breast cancer screening accuracy and efficiency. We aimed to compare the stand-alone performance of an AI system to that of radiologists in detecting breast cancer in DM. METHODS Nine multi-reader, multi-case study datasets previously used for different research purposes in seven countries were collected. Each dataset consisted of DM exams acquired with systems from four different vendors, multiple radiologists' assessments per exam, and ground truth verified by histopathological analysis or follow-up, yielding a total of 2652 exams (653 malignant) and interpretations by 101 radiologists (28 296 independent interpretations). An AI system analyzed these exams yielding a level of suspicion of cancer present between 1 and 10. The detection performance between the radiologists and the AI system was compared using a noninferiority null hypothesis at a margin of 0.05. RESULTS The performance of the AI system was statistically noninferior to that of the average of the 101 radiologists. The AI system had a 0.840 (95% confidence interval [CI] = 0.820 to 0.860) area under the ROC curve and the average of the radiologists was 0.814 (95% CI = 0.787 to 0.841) (difference 95% CI = -0.003 to 0.055). The AI system had an AUC higher than 61.4% of the radiologists. CONCLUSIONS The evaluated AI system achieved a cancer detection accuracy comparable to an average breast radiologist in this retrospective setting. Although promising, the performance and impact of such a system in a screening setting needs further investigation.
Collapse
|
45
|
Badgeley MA, Liu M, Glicksberg BS, Shervey M, Zech J, Shameer K, Lehar J, Oermann EK, McConnell MV, Snyder TM, Dudley JT. CANDI: an R package and Shiny app for annotating radiographs and evaluating computer-aided diagnosis. Bioinformatics 2020; 35:1610-1612. [PMID: 30304439 PMCID: PMC6499410 DOI: 10.1093/bioinformatics/bty855] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2018] [Revised: 08/29/2018] [Accepted: 10/09/2018] [Indexed: 12/05/2022] Open
Abstract
Motivation Radiologists have used algorithms for Computer-Aided Diagnosis (CAD) for decades. These algorithms use machine learning with engineered features, and there have been mixed findings on whether they improve radiologists’ interpretations. Deep learning offers superior performance but requires more training data and has not been evaluated in joint algorithm-radiologist decision systems. Results We developed the Computer-Aided Note and Diagnosis Interface (CANDI) for collaboratively annotating radiographs and evaluating how algorithms alter human interpretation. The annotation app collects classification, segmentation, and image captioning training data, and the evaluation app randomizes the availability of CAD tools to facilitate clinical trials on radiologist enhancement. Availability and implementation Demonstrations and source code are hosted at (https://candi.nextgenhealthcare.org), and (https://github.com/mbadge/candi), respectively, under GPL-3 license. Supplementary information Supplementary material is available at Bioinformatics online.
Collapse
Affiliation(s)
- Marcus A Badgeley
- Department of Genetics and Genomic Sciences, Icahn School of Medicine at Mount Sinai, New York, NY, USA.,Institute for Next Generation Healthcare, Icahn School of Medicine at Mount Sinai, New York, NY, USA.,Verily Life Sciences LLC, South San Francisco, CA, USA
| | - Manway Liu
- Verily Life Sciences LLC, South San Francisco, CA, USA
| | - Benjamin S Glicksberg
- Institute for Computational Health Sciences, University of California, San Francisco, CA, USA
| | - Mark Shervey
- Department of Genetics and Genomic Sciences, Icahn School of Medicine at Mount Sinai, New York, NY, USA.,Institute for Next Generation Healthcare, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| | - John Zech
- Department of Radiology, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| | - Khader Shameer
- Department of Medical Informatics, Northwell Health, Centre for Research Informatics and Innovation, New Hyde Park, NY, USA
| | - Joseph Lehar
- Department of Bioinformatics, Boston University, Boston, MA, USA
| | - Eric K Oermann
- Department of Neurological Surgery, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| | - Michael V McConnell
- Verily Life Sciences LLC, South San Francisco, CA, USA.,Division of Cardiovascular Medicine, Stanford School of Medicine, Stanford, CA, USA
| | | | - Joel T Dudley
- Department of Genetics and Genomic Sciences, Icahn School of Medicine at Mount Sinai, New York, NY, USA.,Institute for Next Generation Healthcare, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| |
Collapse
|
46
|
Chan HP, Samala RK, Hadjiiski LM. CAD and AI for breast cancer-recent development and challenges. Br J Radiol 2020; 93:20190580. [PMID: 31742424 PMCID: PMC7362917 DOI: 10.1259/bjr.20190580] [Citation(s) in RCA: 80] [Impact Index Per Article: 20.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2019] [Revised: 11/13/2019] [Accepted: 11/17/2019] [Indexed: 12/15/2022] Open
Abstract
Computer-aided diagnosis (CAD) has been a popular area of research and development in the past few decades. In CAD, machine learning methods and multidisciplinary knowledge and techniques are used to analyze the patient information and the results can be used to assist clinicians in their decision making process. CAD may analyze imaging information alone or in combination with other clinical data. It may provide the analyzed information directly to the clinician or correlate the analyzed results with the likelihood of certain diseases based on statistical modeling of the past cases in the population. CAD systems can be developed to provide decision support for many applications in the patient care processes, such as lesion detection, characterization, cancer staging, treatment planning and response assessment, recurrence and prognosis prediction. The new state-of-the-art machine learning technique, known as deep learning (DL), has revolutionized speech and text recognition as well as computer vision. The potential of major breakthrough by DL in medical image analysis and other CAD applications for patient care has brought about unprecedented excitement of applying CAD, or artificial intelligence (AI), to medicine in general and to radiology in particular. In this paper, we will provide an overview of the recent developments of CAD using DL in breast imaging and discuss some challenges and practical issues that may impact the advancement of artificial intelligence and its integration into clinical workflow.
Collapse
Affiliation(s)
- Heang-Ping Chan
- Department of Radiology, University of Michigan, Ann Arbor, MI, United States
| | - Ravi K. Samala
- Department of Radiology, University of Michigan, Ann Arbor, MI, United States
| | | |
Collapse
|
47
|
Vollmer S, Mateen BA, Bohner G, Király FJ, Ghani R, Jonsson P, Cumbers S, Jonas A, McAllister KSL, Myles P, Granger D, Birse M, Branson R, Moons KGM, Collins GS, Ioannidis JPA, Holmes C, Hemingway H. Machine learning and artificial intelligence research for patient benefit: 20 critical questions on transparency, replicability, ethics, and effectiveness. BMJ 2020; 368:l6927. [PMID: 32198138 PMCID: PMC11515850 DOI: 10.1136/bmj.l6927] [Citation(s) in RCA: 155] [Impact Index Per Article: 38.8] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 10/22/2019] [Indexed: 12/11/2022]
Affiliation(s)
- Sebastian Vollmer
- Alan Turing Institute, Kings Cross, London, UK
- Departments of Mathematics and Statistics, University of Warwick, Coventry, UK
| | - Bilal A Mateen
- Alan Turing Institute, Kings Cross, London, UK
- Warwick Medical School, University of Warwick, Coventry, UK
- Kings College Hospital, Denmark Hill, London, UK
| | - Gergo Bohner
- Alan Turing Institute, Kings Cross, London, UK
- Departments of Mathematics and Statistics, University of Warwick, Coventry, UK
| | - Franz J Király
- Alan Turing Institute, Kings Cross, London, UK
- Department of Statistical Science, University College London, London, UK
| | | | - Pall Jonsson
- Science Policy and Research, National Institute for Health and Care Excellence, Manchester, UK
| | - Sarah Cumbers
- Health and Social Care Directorate, National Institute for Health and Care Excellence, London, UK
| | - Adrian Jonas
- Data and Analytics Group, National Institute for Health and Care Excellence, London, UK
| | | | - Puja Myles
- Clinical Practice Research Datalink, Medicines and Healthcare products Regulatory Agency, London, UK
| | - David Granger
- Medicines and Healthcare products Regulatory Agency, London, UK
| | - Mark Birse
- Medicines and Healthcare products Regulatory Agency, London, UK
| | - Richard Branson
- Medicines and Healthcare products Regulatory Agency, London, UK
| | - Karel G M Moons
- Julius Centre for Health Sciences and Primary Care, UMC Utrecht, Utrecht University, Utrecht, Netherlands
| | - Gary S Collins
- UK EQUATOR Centre, Centre for Statistics in Medicine, NDORMS, University of Oxford, Oxford, UK
| | - John P A Ioannidis
- Meta-Research Innovation Centre at Stanford, Stanford University, Stanford, CA, USA
| | - Chris Holmes
- Alan Turing Institute, Kings Cross, London, UK
- Department of Statistics, University of Oxford, Oxford OX1 3LB, UK
| | - Harry Hemingway
- Health Data Research UK London, University College London, London, UK
- Institute of Health Informatics, University College London, London, UK
- National Institute for Health Research, University College London Hospitals Biomedical Research Centre, University College London, London, UK
| |
Collapse
|
48
|
Du-Crow E, Astley SM, Hulleman J. Is there a safety-net effect with computer-aided detection? J Med Imaging (Bellingham) 2020; 7:022405. [PMID: 31903408 PMCID: PMC6931663 DOI: 10.1117/1.jmi.7.2.022405] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2019] [Accepted: 12/05/2019] [Indexed: 11/14/2022] Open
Abstract
Computer-aided detection (CAD) systems are used to aid readers interpreting screening mammograms. An expert reader searches the image initially unaided and then once again with the aid of CAD, which prompts automatically detected suspicious regions. This could lead to a "safety-net" effect, where the initial unaided search of the image is adversely affected by the fact that it is preliminary to an additional search with CAD and may, therefore, be less thorough. To investigate the existence of such an effect, we created a visual search experiment for nonexpert observers mirroring breast screening with CAD. Each observer searched 100 images for microcalcification clusters within synthetic images in both prompted (CAD) and unprompted (no-CAD) conditions. Fifty-two participants were recruited for the study, 48 of whom had their eye movements tracked in real-time; the other 4 participants could not be accurately calibrated, so only behavioral data were collected. In the CAD condition, before prompts were displayed, image coverage was significantly lower than coverage in the no-CAD condition (t 47 = 5.29 , p < 0.0001 ). Observer sensitivity was significantly greater for targets marked by CAD than the same targets in the no-CAD condition (t 51 = 6.56 , p < 0.001 ). For targets not marked by CAD, there was no significant difference in observer sensitivity in the CAD condition compared with the same targets in the no-CAD condition (t 51 = 0.54 , p = 0.59 ). These results suggest that the initial search may be influenced by the subsequent availability of CAD; if so, cross-sectional CAD efficacy studies should account for the effect when estimating benefit.
Collapse
Affiliation(s)
- Ethan Du-Crow
- Division of Informatics, Imaging and Data Sciences, University of Manchester, Faculty of Biology, Medicine and Health, Manchester, United Kingdom
| | - Susan M Astley
- Division of Informatics, Imaging and Data Sciences, University of Manchester, Faculty of Biology, Medicine and Health, Manchester, United Kingdom
| | - Johan Hulleman
- Division of Neuroscience and Experimental Psychology, University of Manchester, Faculty of Biology, Medicine and Health, Manchester, United Kingdom
| |
Collapse
|
49
|
Guo K, Yuan M, Wei L, Lu J. Epileptogenic zone localization using a new automatic quantitative analysis based on normal brain glucose metabolism database. Int J Neurosci 2020; 131:128-134. [PMID: 32098541 DOI: 10.1080/00207454.2020.1733561] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
Abstract
OBJECTIVES To assess the clinical value of voxel-based automatic quantitative analysis using a normal brain glucose metabolism database in the preoperative localization of focal intractable temporal lobe epilepsy patients. METHODS Patients with refractory temporal lobe epilepsy who underwent 18F-fluorodeoxyglucose positron emission tomography (18F-FDG PET) imaging were retrospectively enrolled from January to June 2017. Visual analysis was performed by two nuclear medicine radiologists, and the automatic quantitative analysis was carried out using MIMneuro software based the age- and gender-stratified normal brain glucose metabolism database. Setting postoperative outcomes as reference, the consistency between visual analysis and automatic quantitative analysis was tested by Cohen's kappa coefficient, and differences in localization of epileptic foci of the two methods were compared by Chi-square test. RESULTS A total of 32 patients intractable temporal lobe epilepsy were included in this study. There was a moderate agreement between the automatic quantitative analysis based on MIMneuro software and visual analysis (kappa coefficient = 0.472, p = 0.002). In terms of the efficiency of focus localization, the voxel-based automatic quantitative analysis was higher than that of visual analysis (Chi-square value = 6.969, p = 0.008). CONCLUSIONS The voxel-based automatic quantitative analysis combined with normal brain glucose metabolism database had a certain clinical application value for detection temporal lobe epilepsy.
Collapse
Affiliation(s)
- Kun Guo
- Department of Nuclear Medicine, Xuanwu Hospital, Capital Medical University, Beijing, China
| | - Menghui Yuan
- Department of Nuclear Medicine, The Second Affiliated Hospital of Air Force Medical University, Xi'an, Shanxi, China
| | - Longxiao Wei
- Department of Nuclear Medicine, The Second Affiliated Hospital of Air Force Medical University, Xi'an, Shanxi, China
| | - Jie Lu
- Department of Nuclear Medicine, Xuanwu Hospital, Capital Medical University, Beijing, China.,Department of Radiology, Xuanwu Hospital, Capital Medical University, Beijing, China.,Beijing Key Laboratory of Magnetic Resonance Imaging and Brain Informatics, Beijing, China
| |
Collapse
|
50
|
Knippa EE, Berg E, Richard S, Lin Y, Roy Choudhury K, Samei E, Baker JA. Impact of Colorized Display of Mammograms on Lesion Detection. JOURNAL OF BREAST IMAGING 2020; 2:22-28. [PMID: 38424995 DOI: 10.1093/jbi/wbz075] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2019] [Accepted: 10/18/2019] [Indexed: 03/02/2024]
Abstract
OBJECTIVE To assess the effect of the colorized display of digital mammograms on observer detection of subtle breast lesions. METHODS Three separate observer studies compared detection performance using grayscale versus color display of 1) low-contrast mass-like objects in a standardized mammography phantom; 2) simulated microcalcifications in a background of normal breast parenchyma; and 3) standard-of-care clinical digital mammograms with subtle calcifications and masses. Colorization of the images was done by displaying each image pixel in blue, green, and red hues, or gray, maintaining DICOM-calibrated luminance scale and consistent luminance range. For the simulated calcifications and clinical mammogram studies, comparison of detection rates was computed using McNemar's test for paired differences. RESULTS For the phantom study, mass-like object detection was significantly better using a green colormap than grayscale (73.3% vs 70.8%, P = .009), with no significant improvement using blue or red colormaps (72.6% and 72.5%, respectively). For simulated microcalcifications, no significant difference was noted in detection using the green colormap, as compared with grayscale. For clinical digital screening mammograms, no significant difference was noted between gray and green colormaps for detection of microcalcifications. Green color display, however, resulted in decreased sensitivity for detection of subtle masses (63% vs 69%, P = .03). CONCLUSION Although modest improvement was demonstrated for a detection task using colorized display of a standard mammography phantom, no significant improvement was demonstrated using a color display for a simulated clinical detection task, and actual clinical performance was worse for colorized display of mammograms in comparison to standard grayscale display.
Collapse
Affiliation(s)
- Emily E Knippa
- University of Texas Southwestern Medical Center, Department of Radiology, Dallas, TX
| | - Erica Berg
- Wisconsin Radiology Specialists, Milwaukee, WI
| | | | - Yuan Lin
- OPPO US Research Center, AI Department, Palo Alto, CA
| | | | - Ehsan Samei
- Duke University, Departments of Radiology, Physics, and Biomedical Engineering, and Electrical and Computer Engineering, Durham, NC
| | - Jay A Baker
- Duke University Hospital, Department of Radiology, Durham, NC
| |
Collapse
|