101
|
Makino T, Jastrzębski S, Oleszkiewicz W, Chacko C, Ehrenpreis R, Samreen N, Chhor C, Kim E, Lee J, Pysarenko K, Reig B, Toth H, Awal D, Du L, Kim A, Park J, Sodickson DK, Heacock L, Moy L, Cho K, Geras KJ. Differences between human and machine perception in medical diagnosis. Sci Rep 2022; 12:6877. [PMID: 35477730 PMCID: PMC9046399 DOI: 10.1038/s41598-022-10526-z] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2021] [Accepted: 04/06/2022] [Indexed: 02/07/2023] Open
Abstract
Deep neural networks (DNNs) show promise in image-based medical diagnosis, but cannot be fully trusted since they can fail for reasons unrelated to underlying pathology. Humans are less likely to make such superficial mistakes, since they use features that are grounded on medical science. It is therefore important to know whether DNNs use different features than humans. Towards this end, we propose a framework for comparing human and machine perception in medical diagnosis. We frame the comparison in terms of perturbation robustness, and mitigate Simpson's paradox by performing a subgroup analysis. The framework is demonstrated with a case study in breast cancer screening, where we separately analyze microcalcifications and soft tissue lesions. While it is inconclusive whether humans and DNNs use different features to detect microcalcifications, we find that for soft tissue lesions, DNNs rely on high frequency components ignored by radiologists. Moreover, these features are located outside of the region of the images found most suspicious by radiologists. This difference between humans and machines was only visible through subgroup analysis, which highlights the importance of incorporating medical domain knowledge into the comparison.
Collapse
Affiliation(s)
- Taro Makino
- Center for Data Science, New York University, New York, NY, USA. .,Department of Radiology, NYU Langone Health, New York, NY, USA.
| | - Stanisław Jastrzębski
- Center for Data Science, New York University, New York, NY, USA.,Department of Radiology, NYU Langone Health, New York, NY, USA.,Center for Advanced Imaging Innovation and Research, NYU Langone Health, New York, NY, USA
| | - Witold Oleszkiewicz
- Faculty of Electronics and Information Technology, Warsaw University of Technology, Warszawa, Poland
| | - Celin Chacko
- Department of Radiology, NYU Langone Health, New York, NY, USA
| | | | - Naziya Samreen
- Department of Radiology, NYU Langone Health, New York, NY, USA
| | - Chloe Chhor
- Department of Radiology, NYU Langone Health, New York, NY, USA
| | - Eric Kim
- Department of Radiology, NYU Langone Health, New York, NY, USA
| | - Jiyon Lee
- Department of Radiology, NYU Langone Health, New York, NY, USA
| | | | - Beatriu Reig
- Department of Radiology, NYU Langone Health, New York, NY, USA.,Perlmutter Cancer Center, NYU Langone Health, New York, NY, USA
| | - Hildegard Toth
- Department of Radiology, NYU Langone Health, New York, NY, USA.,Perlmutter Cancer Center, NYU Langone Health, New York, NY, USA
| | - Divya Awal
- Department of Radiology, NYU Langone Health, New York, NY, USA
| | - Linda Du
- Department of Radiology, NYU Langone Health, New York, NY, USA
| | - Alice Kim
- Department of Radiology, NYU Langone Health, New York, NY, USA
| | - James Park
- Department of Radiology, NYU Langone Health, New York, NY, USA
| | - Daniel K Sodickson
- Department of Radiology, NYU Langone Health, New York, NY, USA.,Center for Advanced Imaging Innovation and Research, NYU Langone Health, New York, NY, USA.,Vilcek Institute of Graduate Biomedical Sciences, NYU Grossman School of Medicine, New York, NY, USA.,Perlmutter Cancer Center, NYU Langone Health, New York, NY, USA
| | - Laura Heacock
- Department of Radiology, NYU Langone Health, New York, NY, USA.,Perlmutter Cancer Center, NYU Langone Health, New York, NY, USA
| | - Linda Moy
- Department of Radiology, NYU Langone Health, New York, NY, USA.,Center for Advanced Imaging Innovation and Research, NYU Langone Health, New York, NY, USA.,Vilcek Institute of Graduate Biomedical Sciences, NYU Grossman School of Medicine, New York, NY, USA.,Perlmutter Cancer Center, NYU Langone Health, New York, NY, USA
| | - Kyunghyun Cho
- Center for Data Science, New York University, New York, NY, USA.,Department of Computer Science, Courant Institute, New York University, New York, NY, USA
| | - Krzysztof J Geras
- Center for Data Science, New York University, New York, NY, USA. .,Department of Radiology, NYU Langone Health, New York, NY, USA. .,Center for Advanced Imaging Innovation and Research, NYU Langone Health, New York, NY, USA. .,Vilcek Institute of Graduate Biomedical Sciences, NYU Grossman School of Medicine, New York, NY, USA.
| |
Collapse
|
102
|
Vearrier L, Derse AR, Basford JB, Larkin GL, Moskop JC. Artificial Intelligence in Emergency Medicine: Benefits, Risks, and Recommendations. J Emerg Med 2022; 62:492-499. [PMID: 35164977 DOI: 10.1016/j.jemermed.2022.01.001] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2021] [Revised: 12/12/2021] [Accepted: 01/16/2022] [Indexed: 01/04/2023]
Abstract
BACKGROUND Artificial intelligence (AI) can be described as the use of computers to perform tasks that formerly required human cognition. The American Medical Association prefers the term 'augmented intelligence' over 'artificial intelligence' to emphasize the assistive role of computers in enhancing physician skills as opposed to replacing them. The integration of AI into emergency medicine, and clinical practice at large, has increased in recent years, and that trend is likely to continue. DISCUSSION AI has demonstrated substantial potential benefit for physicians and patients. These benefits are transforming the therapeutic relationship from the traditional physician-patient dyad into a triadic doctor-patient-machine relationship. New AI technologies, however, require careful vetting, legal standards, patient safeguards, and provider education. Emergency physicians (EPs) should recognize the limits and risks of AI as well as its potential benefits. CONCLUSIONS EPs must learn to partner with, not capitulate to, AI. AI has proven to be superior to, or on a par with, certain physician skills, such as interpreting radiographs and making diagnoses based on visual cues, such as skin cancer. AI can provide cognitive assistance, but EPs must interpret AI results within the clinical context of individual patients. They must also advocate for patient confidentiality, professional liability coverage, and the essential role of specialty-trained EPs.
Collapse
Affiliation(s)
- Laura Vearrier
- Department of Emergency Medicine, University of Mississippi Medical Center, Jackson, Mississippi
| | - Arthur R Derse
- Center for Bioethics, Medical Humanities, and Department of Emergency Medicine, Medical College of Wisconsin, Wauwatosa, Wisconsin
| | - Jesse B Basford
- Departments of Family and Emergency Medicine, Alabama College of Osteopathic Medicine, Dothan, Alabama
| | - Gregory Luke Larkin
- Department of Emergency Medicine, Northeast Ohio Medical University, Rootstown, Ohio
| | - John C Moskop
- Department of Internal Medicine, Wake Forest School of Medicine, Winston-Salem, North Carolina
| |
Collapse
|
103
|
Ben M'Barek I, Jauvion G, Ceccaldi PF. [Artificial Intelligence in medicine: What about gynecology-obstetric?]. GYNECOLOGIE, OBSTETRIQUE, FERTILITE & SENOLOGIE 2022; 50:340-343. [PMID: 35183787 DOI: 10.1016/j.gofs.2022.02.075] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/06/2021] [Revised: 01/17/2022] [Accepted: 02/10/2022] [Indexed: 06/14/2023]
Affiliation(s)
- I Ben M'Barek
- Service de gynécologie obstétrique, Assistance publique-Hôpitaux de Paris-Beaujon, 100, boulevard du Général-Leclerc, Clichy, France; Université de Paris, 75006 Paris, France; Département de simulation en Santé, Université de Paris, Paris, France.
| | | | - P-F Ceccaldi
- Service de gynécologie obstétrique, Assistance publique-Hôpitaux de Paris-Beaujon, 100, boulevard du Général-Leclerc, Clichy, France; Université de Paris, 75006 Paris, France; Département de simulation en Santé, Université de Paris, Paris, France
| |
Collapse
|
104
|
Wimmer M, Sluiter G, Major D, Lenis D, Berg A, Neubauer T, Buhler K. Multi-Task Fusion for Improving Mammography Screening Data Classification. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:937-950. [PMID: 34788218 DOI: 10.1109/tmi.2021.3129068] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Machine learning and deep learning methods have become essential for computer-assisted prediction in medicine, with a growing number of applications also in the field of mammography. Typically these algorithms are trained for a specific task, e.g., the classification of lesions or the prediction of a mammogram's pathology status. To obtain a comprehensive view of a patient, models which were all trained for the same task(s) are subsequently ensembled or combined. In this work, we propose a pipeline approach, where we first train a set of individual, task-specific models and subsequently investigate the fusion thereof, which is in contrast to the standard model ensembling strategy. We fuse model predictions and high-level features from deep learning models with hybrid patient models to build stronger predictors on patient level. To this end, we propose a multi-branch deep learning model which efficiently fuses features across different tasks and mammograms to obtain a comprehensive patient-level prediction. We train and evaluate our full pipeline on public mammography data, i.e., DDSM and its curated version CBIS-DDSM, and report an AUC score of 0.962 for predicting the presence of any lesion and 0.791 for predicting the presence of malignant lesions on patient level. Overall, our fusion approaches improve AUC scores significantly by up to 0.04 compared to standard model ensembling. Moreover, by providing not only global patient-level predictions but also task-specific model results that are related to radiological features, our pipeline aims to closely support the reading workflow of radiologists.
Collapse
|
105
|
Farina E, Nabhen JJ, Dacoregio MI, Batalini F, Moraes FY. An overview of artificial intelligence in oncology. Future Sci OA 2022; 8:FSO787. [PMID: 35369274 PMCID: PMC8965797 DOI: 10.2144/fsoa-2021-0074] [Citation(s) in RCA: 17] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2021] [Accepted: 01/19/2022] [Indexed: 11/23/2022] Open
Abstract
Cancer is associated with significant morbimortality globally. Advances in screening, diagnosis, management and survivorship were substantial in the last decades, however, challenges in providing personalized and data-oriented care remain. Artificial intelligence (AI), a branch of computer science used for predictions and automation, has emerged as potential solution to improve the healthcare journey and to promote precision in healthcare. AI applications in oncology include, but are not limited to, optimization of cancer research, improvement of clinical practice (eg., prediction of the association of multiple parameters and outcomes - prognosis and response) and better understanding of tumor molecular biology. In this review, we examine the current state of AI in oncology, including fundamentals, current applications, limitations and future perspectives.
Collapse
Affiliation(s)
- Eduardo Farina
- Department of Radiology, Federal University of São Paulo, SP, 04021-001, Brazil; Diagnósticos da America SA (Dasa), 05425-020, Brazil
| | - Jacqueline J Nabhen
- School of Medicine, Federal University of Paraná, Curitiba, PR, 80060-000, Brazil
| | - Maria Inez Dacoregio
- School of Medicine, State University of Centro-Oeste, Guarapuava, PR, 85040-167, Brazil
| | - Felipe Batalini
- Department of Medicine, Division of Medical Oncology, Beth Israel Deaconess Medical Center, Harvard Medical School, Boston, MA 02215, USA
| | - Fabio Y Moraes
- Department of Oncology, Division of Radiation Oncology, Queen's University, Kingston, ON, K7L 3N6, Canada
| |
Collapse
|
106
|
Vuillaume LA, Carval G, Grajoszex M, Ouamara N, Salvestrini JP. Innovation en santé au service des urgences et des cliniciens-chercheurs : le parcours et l’évaluation clinique d’un dispositif médical. ANNALES FRANCAISES DE MEDECINE D URGENCE 2022. [DOI: 10.3166/afmu-2022-0382] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
Abstract
L’alliance du monde de la santé avec celui des nouvelles technologies peut générer un nouveau champ pour la recherche et l’exercice médical. La nouvelle réglementation européenne du marquage Communauté européenne (CE) (2017/745) impose aujourd’hui à tout dispositif médical (DM), y compris les logiciels informatiques, une évaluation clinique rigoureuse. Les services d’urgence sont particulièrement à la pointe en matière d’appétence pour les nouvelles technologies. Ils sont de plus en plus sollicités par des entreprises existantes ou en création, travaillant dans le champ de l’ingénierie à destination du domaine de la santé, mais également par des chercheurs en ingénierie fondamentale. Au regard de ces éléments, il nous semblait pertinent pour les médecins et cliniciens-chercheurs en médecine d’urgence de définir les étapes et jalons d’un projet innovant en santé comprenant un DM. Après avoir passé les premiers stades de développement, matérialisables par l’échelle Technology Readiness Level (TRL), des investigations cliniques sont nécessaires afin d’obtenir un marquage CE auprès d’un organisme notifié. Selon les caractéristiques du DM, différentes modalités de prise en charge par la collectivité sont possibles et peuvent nécessiter des preuves cliniques, médicoéconomiques ou organisationnelles supplémentaires. Les attentes des acteurs et étapes associées au processus d’accès au marché peuvent être un frein au développement des futurs DM. L’association entre chercheurs, cliniciens-chercheurs et industriels devrait permettre l’émergence d’outils novateurs et pratiques pour la médecine d’urgence de demain, à condition d’avoir une bonne connaissance de ces étapes.
Collapse
|
107
|
The Role of Artificial Intelligence in Early Cancer Diagnosis. Cancers (Basel) 2022; 14:cancers14061524. [PMID: 35326674 PMCID: PMC8946688 DOI: 10.3390/cancers14061524] [Citation(s) in RCA: 54] [Impact Index Per Article: 27.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2022] [Revised: 03/08/2022] [Accepted: 03/10/2022] [Indexed: 02/01/2023] Open
Abstract
Improving the proportion of patients diagnosed with early-stage cancer is a key priority of the World Health Organisation. In many tumour groups, screening programmes have led to improvements in survival, but patient selection and risk stratification are key challenges. In addition, there are concerns about limited diagnostic workforces, particularly in light of the COVID-19 pandemic, placing a strain on pathology and radiology services. In this review, we discuss how artificial intelligence algorithms could assist clinicians in (1) screening asymptomatic patients at risk of cancer, (2) investigating and triaging symptomatic patients, and (3) more effectively diagnosing cancer recurrence. We provide an overview of the main artificial intelligence approaches, including historical models such as logistic regression, as well as deep learning and neural networks, and highlight their early diagnosis applications. Many data types are suitable for computational analysis, including electronic healthcare records, diagnostic images, pathology slides and peripheral blood, and we provide examples of how these data can be utilised to diagnose cancer. We also discuss the potential clinical implications for artificial intelligence algorithms, including an overview of models currently used in clinical practice. Finally, we discuss the potential limitations and pitfalls, including ethical concerns, resource demands, data security and reporting standards.
Collapse
|
108
|
Advancements in Oncology with Artificial Intelligence—A Review Article. Cancers (Basel) 2022; 14:cancers14051349. [PMID: 35267657 PMCID: PMC8909088 DOI: 10.3390/cancers14051349] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2022] [Revised: 02/26/2022] [Accepted: 02/28/2022] [Indexed: 02/05/2023] Open
Abstract
Simple Summary With the advancement of artificial intelligence, including machine learning, the field of oncology has seen promising results in cancer detection and classification, epigenetics, drug discovery, and prognostication. In this review, we describe what artificial intelligence is and its function, as well as comprehensively summarize its evolution and role in breast, colorectal, and central nervous system cancers. Understanding the origin and current accomplishments might be essential to improve the quality, accuracy, generalizability, cost-effectiveness, and reliability of artificial intelligence models that can be used in worldwide clinical practice. Students and researchers in the medical field will benefit from a deeper understanding of how to use integrative AI in oncology for innovation and research. Abstract Well-trained machine learning (ML) and artificial intelligence (AI) systems can provide clinicians with therapeutic assistance, potentially increasing efficiency and improving efficacy. ML has demonstrated high accuracy in oncology-related diagnostic imaging, including screening mammography interpretation, colon polyp detection, glioma classification, and grading. By utilizing ML techniques, the manual steps of detecting and segmenting lesions are greatly reduced. ML-based tumor imaging analysis is independent of the experience level of evaluating physicians, and the results are expected to be more standardized and accurate. One of the biggest challenges is its generalizability worldwide. The current detection and screening methods for colon polyps and breast cancer have a vast amount of data, so they are ideal areas for studying the global standardization of artificial intelligence. Central nervous system cancers are rare and have poor prognoses based on current management standards. ML offers the prospect of unraveling undiscovered features from routinely acquired neuroimaging for improving treatment planning, prognostication, monitoring, and response assessment of CNS tumors such as gliomas. By studying AI in such rare cancer types, standard management methods may be improved by augmenting personalized/precision medicine. This review aims to provide clinicians and medical researchers with a basic understanding of how ML works and its role in oncology, especially in breast cancer, colorectal cancer, and primary and metastatic brain cancer. Understanding AI basics, current achievements, and future challenges are crucial in advancing the use of AI in oncology.
Collapse
|
109
|
Retson TA, Hasenstab KA, Kligerman SJ, Jacobs KE, Yen AC, Brouha SS, Hahn LD, Hsiao A. Reader Perceptions and Impact of AI on CT Assessment of Air Trapping. Radiol Artif Intell 2022; 4:e210160. [PMID: 35391767 DOI: 10.1148/ryai.2021210160] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2021] [Revised: 09/22/2021] [Accepted: 10/22/2021] [Indexed: 11/11/2022]
Abstract
Quantitative imaging measurements can be facilitated by artificial intelligence (AI) algorithms, but how they might impact decision-making and be perceived by radiologists remains uncertain. After creation of a dedicated inspiratory-expiratory CT examination and concurrent deployment of a quantitative AI algorithm for assessing air trapping, five cardiothoracic radiologists retrospectively evaluated severity of air trapping on 17 examination studies. Air trapping severity of each lobe was evaluated in three stages: qualitatively (visually); semiquantitatively, allowing manual region-of-interest measurements; and quantitatively, using results from an AI algorithm. Readers were surveyed on each case for their perceptions of the AI algorithm. The algorithm improved interreader agreement (intraclass correlation coefficients: visual, 0.28; semiquantitative, 0.40; quantitative, 0.84; P < .001) and improved correlation with pulmonary function testing (forced expiratory volume in 1 second-to-forced vital capacity ratio) (visual r = -0.26, semiquantitative r = -0.32, quantitative r = -0.44). Readers perceived moderate agreement with the AI algorithm (Likert scale average, 3.7 of 5), a mild impact on their final assessment (average, 2.6), and a neutral perception of overall utility (average, 3.5). Though the AI algorithm objectively improved interreader consistency and correlation with pulmonary function testing, individual readers did not immediately perceive this benefit, revealing a potential barrier to clinical adoption. Keywords: Technology Assessment, Quantification © RSNA, 2021.
Collapse
Affiliation(s)
- Tara A Retson
- Department of Radiology, University of California, San Diego, 9452 Medical Center Dr, 4th Floor, La Jolla, CA 92037 (T.A.R., S.J.K., K.E.J., A.C.Y., S.S.B., L.D.H., A.H.); and Department of Mathematics and Statistics, San Diego State University, San Diego, Calif (K.A.H.)
| | - Kyle A Hasenstab
- Department of Radiology, University of California, San Diego, 9452 Medical Center Dr, 4th Floor, La Jolla, CA 92037 (T.A.R., S.J.K., K.E.J., A.C.Y., S.S.B., L.D.H., A.H.); and Department of Mathematics and Statistics, San Diego State University, San Diego, Calif (K.A.H.)
| | - Seth J Kligerman
- Department of Radiology, University of California, San Diego, 9452 Medical Center Dr, 4th Floor, La Jolla, CA 92037 (T.A.R., S.J.K., K.E.J., A.C.Y., S.S.B., L.D.H., A.H.); and Department of Mathematics and Statistics, San Diego State University, San Diego, Calif (K.A.H.)
| | - Kathleen E Jacobs
- Department of Radiology, University of California, San Diego, 9452 Medical Center Dr, 4th Floor, La Jolla, CA 92037 (T.A.R., S.J.K., K.E.J., A.C.Y., S.S.B., L.D.H., A.H.); and Department of Mathematics and Statistics, San Diego State University, San Diego, Calif (K.A.H.)
| | - Andrew C Yen
- Department of Radiology, University of California, San Diego, 9452 Medical Center Dr, 4th Floor, La Jolla, CA 92037 (T.A.R., S.J.K., K.E.J., A.C.Y., S.S.B., L.D.H., A.H.); and Department of Mathematics and Statistics, San Diego State University, San Diego, Calif (K.A.H.)
| | - Sharon S Brouha
- Department of Radiology, University of California, San Diego, 9452 Medical Center Dr, 4th Floor, La Jolla, CA 92037 (T.A.R., S.J.K., K.E.J., A.C.Y., S.S.B., L.D.H., A.H.); and Department of Mathematics and Statistics, San Diego State University, San Diego, Calif (K.A.H.)
| | - Lewis D Hahn
- Department of Radiology, University of California, San Diego, 9452 Medical Center Dr, 4th Floor, La Jolla, CA 92037 (T.A.R., S.J.K., K.E.J., A.C.Y., S.S.B., L.D.H., A.H.); and Department of Mathematics and Statistics, San Diego State University, San Diego, Calif (K.A.H.)
| | - Albert Hsiao
- Department of Radiology, University of California, San Diego, 9452 Medical Center Dr, 4th Floor, La Jolla, CA 92037 (T.A.R., S.J.K., K.E.J., A.C.Y., S.S.B., L.D.H., A.H.); and Department of Mathematics and Statistics, San Diego State University, San Diego, Calif (K.A.H.)
| |
Collapse
|
110
|
Feng S, Liu Q, Patel A, Bazai SU, Jin C, Kim JS, Sarrafzadeh M, Azzollini D, Yeoh J, Kim E, Gordon S, Jang‐Jaccard J, Urschler M, Barnard S, Fong A, Simmers C, Tarr GP, Wilson B. Automated pneumothorax triaging in chest X‐rays in the New Zealand population using deep‐learning algorithms. J Med Imaging Radiat Oncol 2022; 66:1035-1043. [DOI: 10.1111/1754-9485.13393] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2021] [Accepted: 02/03/2022] [Indexed: 11/30/2022]
Affiliation(s)
- Sijing Feng
- Department of Radiology Dunedin Hospital Dunedin New Zealand
| | - Qixiu Liu
- Counties Manukau Health Auckland New Zealand
| | - Aakash Patel
- Dunedin School of Medicine Dunedin Hospital Dunedin New Zealand
| | - Sibghat Ullah Bazai
- School of Natural and Computational Sciences Massey University Palmerston North New Zealand
| | | | - Ji Soo Kim
- Auckland District Health Board Auckland New Zealand
| | | | | | - Jason Yeoh
- Auckland District Health Board Auckland New Zealand
| | - Eve Kim
- Auckland District Health Board Auckland New Zealand
| | - Simon Gordon
- Waikato District Health Board Hamilton New Zealand
| | - Julian Jang‐Jaccard
- School of Natural and Computational Sciences Massey University Palmerston North New Zealand
| | - Martin Urschler
- School of Computer Science University of Auckland Auckland New Zealand
| | | | - Amy Fong
- Department of Radiology Dunedin Hospital Dunedin New Zealand
| | - Cameron Simmers
- Department of Radiology Dunedin Hospital Dunedin New Zealand
| | | | - Ben Wilson
- Department of Radiology Dunedin Hospital Dunedin New Zealand
| |
Collapse
|
111
|
Deep learning in image-based breast and cervical cancer detection: a systematic review and meta-analysis. NPJ Digit Med 2022; 5:19. [PMID: 35169217 PMCID: PMC8847584 DOI: 10.1038/s41746-022-00559-z] [Citation(s) in RCA: 24] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2021] [Accepted: 12/22/2021] [Indexed: 12/15/2022] Open
Abstract
Accurate early detection of breast and cervical cancer is vital for treatment success. Here, we conduct a meta-analysis to assess the diagnostic performance of deep learning (DL) algorithms for early breast and cervical cancer identification. Four subgroups are also investigated: cancer type (breast or cervical), validation type (internal or external), imaging modalities (mammography, ultrasound, cytology, or colposcopy), and DL algorithms versus clinicians. Thirty-five studies are deemed eligible for systematic review, 20 of which are meta-analyzed, with a pooled sensitivity of 88% (95% CI 85–90%), specificity of 84% (79–87%), and AUC of 0.92 (0.90–0.94). Acceptable diagnostic performance with analogous DL algorithms was highlighted across all subgroups. Therefore, DL algorithms could be useful for detecting breast and cervical cancer using medical imaging, having equivalent performance to human clinicians. However, this tentative assertion is based on studies with relatively poor designs and reporting, which likely caused bias and overestimated algorithm performance. Evidence-based, standardized guidelines around study methods and reporting are required to improve the quality of DL research.
Collapse
|
112
|
Improved automated early detection of breast cancer based on high resolution 3D micro-CT microcalcification images. BMC Cancer 2022; 22:162. [PMID: 35148703 PMCID: PMC8832731 DOI: 10.1186/s12885-021-09133-4] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2021] [Accepted: 12/20/2021] [Indexed: 11/10/2022] Open
Abstract
Background The detection of suspicious microcalcifications on mammography represents one of the earliest signs of a malignant breast tumor. Assessing microcalcifications’ characteristics based on their appearance on 2D breast imaging modalities is in many cases challenging for radiologists. The aims of this study were to: (a) analyse the association of shape and texture properties of breast microcalcifications (extracted by scanning breast tissue with a high resolution 3D scanner) with malignancy, (b) evaluate microcalcifications’ potential to diagnose benign/malignant patients. Methods Biopsy samples of 94 female patients with suspicious microcalcifications detected during a mammography, were scanned using a micro-CT scanner at a resolution of 9 μm. Several preprocessing techniques were applied on 3504 extracted microcalcifications. A high amount of radiomic features were extracted in an attempt to capture differences among microcalcifications occurring in benign and malignant lesions. Machine learning algorithms were used to diagnose: (a) individual microcalcifications, (b) samples. For the samples, several methodologies to combine individual microcalcification results into sample results were evaluated. Results We could classify individual microcalcifications with 77.32% accuracy, 61.15% sensitivity and 89.76% specificity. At the sample level diagnosis, we achieved an accuracy of 84.04%, sensitivity of 86.27% and specificity of 81.39%. Conclusions By studying microcalcifications’ characteristics at a level of details beyond what is currently possible by using conventional breast imaging modalities, our classification results demonstrated a strong association between breast microcalcifications and malignancies. Microcalcification’s texture features extracted in transform domains, have higher discriminating power to classify benign/malignant individual microcalcifications and samples compared to pure shape-features.
Collapse
|
113
|
Anderson AW, Marinovich ML, Houssami N, Lowry KP, Elmore JG, Buist DS, Hofvind S, Lee CI. Independent External Validation of Artificial Intelligence Algorithms for Automated Interpretation of Screening Mammography: A Systematic Review. J Am Coll Radiol 2022; 19:259-273. [PMID: 35065909 PMCID: PMC8857031 DOI: 10.1016/j.jacr.2021.11.008] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2021] [Revised: 11/02/2021] [Accepted: 11/03/2021] [Indexed: 02/03/2023]
Abstract
PURPOSE The aim of this study was to describe the current state of science regarding independent external validation of artificial intelligence (AI) technologies for screening mammography. METHODS A systematic review was performed across five databases (Embase, PubMed, IEEE Explore, Engineer Village, and arXiv) through December 10, 2020. Studies that used screening examinations from real-world settings to externally validate AI algorithms for mammographic cancer detection were included. The main outcome was diagnostic accuracy, defined by area under the receiver operating characteristic curve (AUC). Performance was also compared between radiologists and either stand-alone AI or combined radiologist and AI interpretation. Study quality was assessed using the Quality Assessment of Diagnostic Accuracy Studies 2 tool. RESULTS After data extraction, 13 studies met the inclusion criteria (148,361 total patients). Most studies (77% [n = 10]) evaluated commercially available AI algorithms. Studies included retrospective reader studies (46% [n = 6]), retrospective simulation studies (38% [n = 5]), or both (15% [n = 2]). Across 5 studies comparing stand-alone AI with radiologists, 60% (n = 3) demonstrated improved accuracy with AI (AUC improvement range, 0.02-0.13). All 5 studies comparing combined radiologist and AI interpretation with radiologists alone demonstrated improved accuracy with AI (AUC improvement range, 0.028-0.115). Most studies had risk for bias or applicability concerns for patient selection (69% [n = 9]) and the reference standard (69% [n = 9]). Only two studies obtained ground-truth cancer outcomes through regional cancer registry linkage. CONCLUSIONS To date, external validation efforts for AI screening mammographic technologies suggest small potential diagnostic accuracy improvements but have been retrospective in nature and suffer from risk for bias and applicability concerns.
Collapse
Affiliation(s)
- Anna W. Anderson
- Department of Radiology, University of Washington School of Medicine, Seattle, WA
| | - M. Luke Marinovich
- Curtin School of Population Health, Curtin University, Bentley, Western Australia, Australia
| | - Nehmat Houssami
- The Daffodil Centre, the University of Sydney, a joint venture with Cancer Council NSW, Sydney, New South Wales, Australia
| | - Kathryn P. Lowry
- Department of Radiology, University of Washington School of Medicine, Seattle, WA
| | - Joann G. Elmore
- David Geffen School of Medicine at University of California at Los Angeles, Los Angeles, CA
| | - Diana S.M. Buist
- Kaiser Permanente Washington Health Research Institute, Seattle, WA
| | | | - Christoph I. Lee
- Department of Radiology, University of Washington School of Medicine, Seattle, WA
| |
Collapse
|
114
|
Ram S, Campbell T, Lourenco AP. Online or Offline: Does It Matter? A Review of Existing Interpretation Approaches and Their Effect on Screening Mammography Metrics, Patient Satisfaction, and Cost. JOURNAL OF BREAST IMAGING 2022; 4:3-9. [PMID: 38422414 DOI: 10.1093/jbi/wbab086] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2021] [Indexed: 03/02/2024]
Abstract
The ideal practice routine for screening mammography would optimize performance metrics and minimize costs, while also maximizing patient satisfaction. The main approaches to screening mammography interpretation include batch offline, non-batch offline, interrupted online, and uninterrupted online reading, each of which has its own advantages and drawbacks. This article reviews the current literature on approaches to screening mammography interpretation, potential effects of newer technologies, and promising artificial intelligence resources that could improve workflow efficiency in the future.
Collapse
Affiliation(s)
- Shruthi Ram
- Alpert Medical School of Brown University and Rhode Island Hospital, Department of Diagnostic Imaging, Providence, RI, USA
| | - Tyler Campbell
- Alpert Medical School of Brown University and Rhode Island Hospital, Department of Diagnostic Imaging, Providence, RI, USA
| | - Ana P Lourenco
- Alpert Medical School of Brown University and Rhode Island Hospital, Department of Diagnostic Imaging, Providence, RI, USA
| |
Collapse
|
115
|
Ito Y, Miyoshi A, Ueda Y, Tanaka Y, Nakae R, Morimoto A, Shiomi M, Enomoto T, Sekine M, Sasagawa T, Yoshino K, Harada H, Nakamura T, Murata T, Hiramatsu K, Saito J, Yagi J, Tanaka Y, Kimura T. An artificial intelligence-assisted diagnostic system improves the accuracy of image diagnosis of uterine cervical lesions. Mol Clin Oncol 2022; 16:27. [PMID: 34987798 DOI: 10.3892/mco.2021.2460] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2021] [Accepted: 10/07/2021] [Indexed: 12/31/2022] Open
Abstract
The present study created an artificial intelligence (AI)-automated diagnostics system for uterine cervical lesions and assessed the performance of these images for AI diagnostic imaging of pathological cervical lesions. A total of 463 colposcopic images were analyzed. The traditional colposcopy diagnoses were compared to those obtained by AI image diagnosis. Next, 100 images were presented to a panel of 32 gynecologists who independently examined each image in a blinded fashion and diagnosed them for four categories of tumors. Then, the 32 gynecologists revisited their diagnosis for each image after being informed of the AI diagnosis. The present study assessed any changes in physician diagnosis and the accuracy of AI-image-assisted diagnosis (AISD). The accuracy of AI was 57.8% for normal, 35.4% for cervical intraepithelial neoplasia (CIN)1, 40.5% for CIN2-3 and 44.2% for invasive cancer. The accuracy of gynecologist diagnoses from cervical pathological images, before knowing the AI image diagnosis, was 54.4% for CIN2-3 and 38.9% for invasive cancer. After learning of the AISD, their accuracy improved to 58.0% for CIN2-3 and 48.5% for invasive cancer. AI-assisted image diagnosis was able to improve gynecologist diagnosis accuracy significantly (P<0.01) for invasive cancer and tended to improve their accuracy for CIN2-3 (P=0.14).
Collapse
Affiliation(s)
- Yu Ito
- Department of Obstetrics and Gynecology, Osaka University Graduate School of Medicine, Suita, Osaka 567-0871, Japan
| | - Ai Miyoshi
- Department of Obstetrics and Gynecology, Osaka University Graduate School of Medicine, Suita, Osaka 567-0871, Japan
| | - Yutaka Ueda
- Department of Obstetrics and Gynecology, Osaka University Graduate School of Medicine, Suita, Osaka 567-0871, Japan
| | - Yusuke Tanaka
- Department of Obstetrics and Gynecology, Osaka University Graduate School of Medicine, Suita, Osaka 567-0871, Japan
| | - Ruriko Nakae
- Department of Obstetrics and Gynecology, Osaka University Graduate School of Medicine, Suita, Osaka 567-0871, Japan
| | - Akiko Morimoto
- Department of Obstetrics and Gynecology, Osaka University Graduate School of Medicine, Suita, Osaka 567-0871, Japan
| | - Mayu Shiomi
- Department of Obstetrics and Gynecology, Osaka University Graduate School of Medicine, Suita, Osaka 567-0871, Japan
| | - Takayuki Enomoto
- Department of Obstetrics and Gynecology, Niigata University Graduate School of Medicine, Chuo-ku, Niigata 951-8520, Japan
| | - Masayuki Sekine
- Department of Obstetrics and Gynecology, Niigata University Graduate School of Medicine, Chuo-ku, Niigata 951-8520, Japan
| | - Toshiyuki Sasagawa
- Department of Obstetrics and Gynecology, Kanazawa Medical University, Uchinada, Ishikawa 920-0293, Japan
| | - Kiyoshi Yoshino
- Department of Obstetrics and Gynecology, University of Occupational and Environmental Health, Kitakyushu, Fukuoka 807-8556, Japan
| | - Hiroshi Harada
- Department of Obstetrics and Gynecology, University of Occupational and Environmental Health, Kitakyushu, Fukuoka 807-8556, Japan
| | - Takafumi Nakamura
- Department of Obstetrics and Gynecology, Kawasaki Medical University, Kurashiki, Okayama 701-0192, Japan
| | - Takuya Murata
- Department of Obstetrics and Gynecology, Kawasaki Medical University, Kurashiki, Okayama 701-0192, Japan
| | - Keizo Hiramatsu
- Hiramatsu Obstetrics and Gynecology Clinic, Kishiwada-shi, Osaka 583-0024, Japan
| | - Junko Saito
- Saito Women Clinic, Yodogawa-ku, Osaka 532-0003, Japan
| | - Junko Yagi
- Ladies Clinic Yagi, Senboku-gunn, Osaka 595-0805, Japan
| | | | - Tadashi Kimura
- Department of Obstetrics and Gynecology, Osaka University Graduate School of Medicine, Suita, Osaka 567-0871, Japan
| |
Collapse
|
116
|
Hickman SE, Woitek R, Le EPV, Im YR, Mouritsen Luxhøj C, Aviles-Rivero AI, Baxter GC, MacKay JW, Gilbert FJ. Machine Learning for Workflow Applications in Screening Mammography: Systematic Review and Meta-Analysis. Radiology 2022; 302:88-104. [PMID: 34665034 PMCID: PMC8717814 DOI: 10.1148/radiol.2021210391] [Citation(s) in RCA: 38] [Impact Index Per Article: 19.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2021] [Revised: 07/14/2021] [Accepted: 08/05/2021] [Indexed: 01/03/2023]
Abstract
Background Advances in computer processing and improvements in data availability have led to the development of machine learning (ML) techniques for mammographic imaging. Purpose To evaluate the reported performance of stand-alone ML applications for screening mammography workflow. Materials and Methods Ovid Embase, Ovid Medline, Cochrane Central Register of Controlled Trials, Scopus, and Web of Science literature databases were searched for relevant studies published from January 2012 to September 2020. The study was registered with the PROSPERO International Prospective Register of Systematic Reviews (protocol no. CRD42019156016). Stand-alone technology was defined as a ML algorithm that can be used independently of a human reader. Studies were quality assessed using the Quality Assessment of Diagnostic Accuracy Studies 2 and the Prediction Model Risk of Bias Assessment Tool, and reporting was evaluated using the Checklist for Artificial Intelligence in Medical Imaging. A primary meta-analysis included the top-performing algorithm and corresponding reader performance from which pooled summary estimates for the area under the receiver operating characteristic curve (AUC) were calculated using a bivariate model. Results Fourteen articles were included, which detailed 15 studies for stand-alone detection (n = 8) and triage (n = 7). Triage studies reported that 17%-91% of normal mammograms identified could be read by adapted screening, while "missing" an estimated 0%-7% of cancers. In total, an estimated 185 252 cases from three countries with more than 39 readers were included in the primary meta-analysis. The pooled sensitivity, specificity, and AUC was 75.4% (95% CI: 65.6, 83.2; P = .11), 90.6% (95% CI: 82.9, 95.0; P = .40), and 0.89 (95% CI: 0.84, 0.98), respectively, for algorithms, and 73.0% (95% CI: 60.7, 82.6), 88.6% (95% CI: 72.4, 95.8), and 0.85 (95% CI: 0.78, 0.97), respectively, for readers. Conclusion Machine learning (ML) algorithms that demonstrate a stand-alone application in mammographic screening workflows achieve or even exceed human reader detection performance and improve efficiency. However, this evidence is from a small number of retrospective studies. Therefore, further rigorous independent external prospective testing of ML algorithms to assess performance at preassigned thresholds is required to support these claims. ©RSNA, 2021 Online supplemental material is available for this article. See also the editorial by Whitman and Moseley in this issue.
Collapse
Affiliation(s)
- Sarah E. Hickman
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Ramona Woitek
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Elizabeth Phuong Vi Le
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Yu Ri Im
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Carina Mouritsen Luxhøj
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Angelica I. Aviles-Rivero
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Gabrielle C. Baxter
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - James W. MacKay
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| | - Fiona J. Gilbert
- From the Department of Radiology (S.E.H., R.W., G.C.B., J.W.M.,
F.J.G.) and Department of Medicine (E.P.V.L., Y.R.I., C.M.L.), University of
Cambridge School of Clinical Medicine, Box 218, Cambridge Biomedical Campus,
Cambridge, CB2 0QQ, England; Department of Radiology, Addenbrooke's
Hospital, Cambridge University Hospitals National Health Service Foundation
Trust, Cambridge, England (R.W., F.J.G.); Department of Biomedical Imaging and
Image-guided Therapy, Medical University of Vienna, Vienna, Austria (R.W.);
Department of Pure Mathematics and Mathematical Statistics, University of
Cambridge, Cambridge, England (A.I.A.R.); and Norwich Medical School, University
of East Anglia, Norwich, England (J.W.M.)
| |
Collapse
|
117
|
Mueller B, Kinoshita T, Peebles A, Graber MA, Lee S. Artificial intelligence and machine learning in emergency medicine: a narrative review. Acute Med Surg 2022; 9:e740. [PMID: 35251669 PMCID: PMC8887797 DOI: 10.1002/ams2.740] [Citation(s) in RCA: 20] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2021] [Revised: 01/26/2022] [Accepted: 02/06/2022] [Indexed: 12/20/2022] Open
Abstract
AIM The emergence and evolution of artificial intelligence (AI) has generated increasing interest in machine learning applications for health care. Specifically, researchers are grasping the potential of machine learning solutions to enhance the quality of care in emergency medicine. METHODS We undertook a narrative review of published works on machine learning applications in emergency medicine and provide a synopsis of recent developments. RESULTS This review describes fundamental concepts of machine learning and presents clinical applications for triage, risk stratification specific to disease, medical imaging, and emergency department operations. Additionally, we consider how machine learning models could contribute to the improvement of causal inference in medicine, and to conclude, we discuss barriers to safe implementation of AI. CONCLUSION We intend that this review serves as an introduction to AI and machine learning in emergency medicine.
Collapse
Affiliation(s)
- Brianna Mueller
- Department of Business Analytics The University of Iowa Tippie College of Business Iowa City Iowa USA
| | | | - Alexander Peebles
- Department of Emergency Medicine The University of Iowa Carver College of Medicine Iowa City Iowa USA
| | - Mark A Graber
- Department of Emergency Medicine The University of Iowa Carver College of Medicine Iowa City Iowa USA
| | - Sangil Lee
- Department of Emergency Medicine The University of Iowa Carver College of Medicine Iowa City Iowa USA
| |
Collapse
|
118
|
Lee JH, Kim KH, Lee EH, Ahn JS, Ryu JK, Park YM, Shin GW, Kim YJ, Choi HY. Improving the Performance of Radiologists Using Artificial Intelligence-Based Detection Support Software for Mammography: A Multi-Reader Study. Korean J Radiol 2022; 23:505-516. [PMID: 35434976 PMCID: PMC9081685 DOI: 10.3348/kjr.2021.0476] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2021] [Revised: 01/04/2022] [Accepted: 01/24/2022] [Indexed: 12/24/2022] Open
Abstract
Objective To evaluate whether artificial intelligence (AI) for detecting breast cancer on mammography can improve the performance and time efficiency of radiologists reading mammograms. Materials and Methods A commercial deep learning-based software for mammography was validated using external data collected from 200 patients, 100 each with and without breast cancer (40 with benign lesions and 60 without lesions) from one hospital. Ten readers, including five breast specialist radiologists (BSRs) and five general radiologists (GRs), assessed all mammography images using a seven-point scale to rate the likelihood of malignancy in two sessions, with and without the aid of the AI-based software, and the reading time was automatically recorded using a web-based reporting system. Two reading sessions were conducted with a two-month washout period in between. Differences in the area under the receiver operating characteristic curve (AUROC), sensitivity, specificity, and reading time between reading with and without AI were analyzed, accounting for data clustering by readers when indicated. Results The AUROC of the AI alone, BSR (average across five readers), and GR (average across five readers) groups was 0.915 (95% confidence interval, 0.876–0.954), 0.813 (0.756–0.870), and 0.684 (0.616–0.752), respectively. With AI assistance, the AUROC significantly increased to 0.884 (0.840–0.928) and 0.833 (0.779–0.887) in the BSR and GR groups, respectively (p = 0.007 and p < 0.001, respectively). Sensitivity was improved by AI assistance in both groups (74.6% vs. 88.6% in BSR, p < 0.001; 52.1% vs. 79.4% in GR, p < 0.001), but the specificity did not differ significantly (66.6% vs. 66.4% in BSR, p = 0.238; 70.8% vs. 70.0% in GR, p = 0.689). The average reading time pooled across readers was significantly decreased by AI assistance for BSRs (82.73 vs. 73.04 seconds, p < 0.001) but increased in GRs (35.44 vs. 42.52 seconds, p < 0.001). Conclusion AI-based software improved the performance of radiologists regardless of their experience and affected the reading time.
Collapse
Affiliation(s)
| | | | - Eun Hye Lee
- Department of Radiology, Soonchunhyang University Bucheon Hospital, Soonchunhyang University College of Medicine, Bucheon, Korea
| | | | - Jung Kyu Ryu
- Department of Radiology, Kyung Hee University Hospital at Gangdong, Seoul, Korea
| | - Young Mi Park
- Department of Radiology, Inje University Busan Paik Hospital, Inje University College of Medicine, Busan, Korea
| | - Gi Won Shin
- Department of Radiology, Inje University Busan Paik Hospital, Inje University College of Medicine, Busan, Korea
| | - Young Joong Kim
- Department of Radiology, Konyang University Hospital, Konyang University College of Medicine, Daejeon, Korea
| | - Hye Young Choi
- Department of Radiology, Gyeongsang National University Hospital, Jinju, Korea
| |
Collapse
|
119
|
Dee EC, Yu RC, Celi LA, Nehal US. AIM and Business Models of Healthcare. Artif Intell Med 2022. [DOI: 10.1007/978-3-030-64573-1_247] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
120
|
Cè M, Caloro E, Pellegrino ME, Basile M, Sorce A, Fazzini D, Oliva G, Cellina M. Artificial intelligence in breast cancer imaging: risk stratification, lesion detection and classification, treatment planning and prognosis-a narrative review. EXPLORATION OF TARGETED ANTI-TUMOR THERAPY 2022; 3:795-816. [PMID: 36654817 PMCID: PMC9834285 DOI: 10.37349/etat.2022.00113] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2022] [Accepted: 09/28/2022] [Indexed: 12/28/2022] Open
Abstract
The advent of artificial intelligence (AI) represents a real game changer in today's landscape of breast cancer imaging. Several innovative AI-based tools have been developed and validated in recent years that promise to accelerate the goal of real patient-tailored management. Numerous studies confirm that proper integration of AI into existing clinical workflows could bring significant benefits to women, radiologists, and healthcare systems. The AI-based approach has proved particularly useful for developing new risk prediction models that integrate multi-data streams for planning individualized screening protocols. Furthermore, AI models could help radiologists in the pre-screening and lesion detection phase, increasing diagnostic accuracy, while reducing workload and complications related to overdiagnosis. Radiomics and radiogenomics approaches could extrapolate the so-called imaging signature of the tumor to plan a targeted treatment. The main challenges to the development of AI tools are the huge amounts of high-quality data required to train and validate these models and the need for a multidisciplinary team with solid machine-learning skills. The purpose of this article is to present a summary of the most important AI applications in breast cancer imaging, analyzing possible challenges and new perspectives related to the widespread adoption of these new tools.
Collapse
Affiliation(s)
- Maurizio Cè
- Postgraduate School in Diagnostic and Interventional Radiology, University of Milan, 20122 Milan, Italy,Correspondence: Maurizio Cè, Postgraduate School in Diagnostic and Interventional Radiology, University of Milan, Via Festa del Perdono, 7, 20122 Milan, Italy.
| | - Elena Caloro
- Postgraduate School in Diagnostic and Interventional Radiology, University of Milan, 20122 Milan, Italy
| | - Maria E. Pellegrino
- Postgraduate School in Diagnostic and Interventional Radiology, University of Milan, 20122 Milan, Italy
| | - Mariachiara Basile
- Postgraduate School in Diagnostic and Interventional Radiology, University of Milan, 20122 Milan, Italy
| | - Adriana Sorce
- Postgraduate School in Diagnostic and Interventional Radiology, University of Milan, 20122 Milan, Italy
| | | | - Giancarlo Oliva
- Department of Radiology, ASST Fatebenefratelli Sacco, 20121 Milan, Italy
| | - Michaela Cellina
- Department of Radiology, ASST Fatebenefratelli Sacco, 20121 Milan, Italy
| |
Collapse
|
121
|
Liu Q, Qu M, Sun L, Wang H. Accuracy of ultrasonic artificial intelligence in diagnosing benign and malignant breast diseases: A protocol for systematic review and meta-analysis. Medicine (Baltimore) 2021; 100:e28289. [PMID: 34918704 PMCID: PMC8678017 DOI: 10.1097/md.0000000000028289] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/24/2021] [Accepted: 11/29/2021] [Indexed: 01/05/2023] Open
Abstract
BACKGROUND Artificial intelligence system is a deep learning system based on computer-assisted ultrasonic image diagnosis, which can extract morphological features of breast mass and conduct objective and efficient image analysis, thus automatically intelligent classification of breast mass, avoiding subjective error and improving the accuracy of diagnosis.[1-2] A large number of studies have confirmed that artificial intelligence (AI) has high effectiveness and reliability in the differential diagnosis of benign and malignant breast diseases.[3-4] However, the results of these studies have been contradictory. Therefore, this meta-analysis tested the hypothesis that artificial intelligence system is accurate in distinguishing benign and malignant breast diseases. METHODS We will search PubMed, Web of Science, Cochrane Library, and Chinese biomedical databases from their inceptions to the November 20, 2021, without language restrictions. Two authors will independently carry out searching literature records, scanning titles and abstracts, full texts, collecting data, and assessing risk of bias. Review Manager 5.2 and Stata14.0 software will be used for data analysis. RESULTS This systematic review will determine the accuracy of AI in the differential diagnosis of benign and malignant breast diseases. CONCLUSION Its findings will provide helpful evidence for the accuracy of AI in the differential diagnosis of benign and malignant breast diseases. SYSTEMATIC REVIEW REGISTRATION INPLASY2021110087.
Collapse
|
122
|
A case-based interpretable deep learning model for classification of mass lesions in digital mammography. NAT MACH INTELL 2021. [DOI: 10.1038/s42256-021-00423-x] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
123
|
Decuyper M, Maebe J, Van Holen R, Vandenberghe S. Artificial intelligence with deep learning in nuclear medicine and radiology. EJNMMI Phys 2021; 8:81. [PMID: 34897550 PMCID: PMC8665861 DOI: 10.1186/s40658-021-00426-y] [Citation(s) in RCA: 21] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2021] [Accepted: 11/19/2021] [Indexed: 12/19/2022] Open
Abstract
The use of deep learning in medical imaging has increased rapidly over the past few years, finding applications throughout the entire radiology pipeline, from improved scanner performance to automatic disease detection and diagnosis. These advancements have resulted in a wide variety of deep learning approaches being developed, solving unique challenges for various imaging modalities. This paper provides a review on these developments from a technical point of view, categorizing the different methodologies and summarizing their implementation. We provide an introduction to the design of neural networks and their training procedure, after which we take an extended look at their uses in medical imaging. We cover the different sections of the radiology pipeline, highlighting some influential works and discussing the merits and limitations of deep learning approaches compared to other traditional methods. As such, this review is intended to provide a broad yet concise overview for the interested reader, facilitating adoption and interdisciplinary research of deep learning in the field of medical imaging.
Collapse
Affiliation(s)
- Milan Decuyper
- Department of Electronics and Information Systems, Ghent University, Ghent, Belgium
| | - Jens Maebe
- Department of Electronics and Information Systems, Ghent University, Ghent, Belgium
| | - Roel Van Holen
- Department of Electronics and Information Systems, Ghent University, Ghent, Belgium
| | - Stefaan Vandenberghe
- Department of Electronics and Information Systems, Ghent University, Ghent, Belgium
| |
Collapse
|
124
|
Feng S, Azzollini D, Kim JS, Jin CK, Gordon SP, Yeoh J, Kim E, Han M, Lee A, Patel A, Wu J, Urschler M, Fong A, Simmers C, Tarr GP, Barnard S, Wilson B. Curation of the CANDID-PTX Dataset with Free-Text Reports. Radiol Artif Intell 2021; 3:e210136. [PMID: 34870223 DOI: 10.1148/ryai.2021210136] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2021] [Revised: 09/20/2021] [Accepted: 09/27/2021] [Indexed: 12/22/2022]
Abstract
Supplemental material is available for this article. Keywords: Conventional Radiography, Thorax, Trauma, Ribs, Catheters, Segmentation, Diagnosis, Classification, Supervised Learning, Machine Learning © RSNA, 2021.
Collapse
Affiliation(s)
- Sijing Feng
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Damian Azzollini
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Ji Soo Kim
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Cheng-Kai Jin
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Simon P Gordon
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Jason Yeoh
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Eve Kim
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Mina Han
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Andrew Lee
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Aakash Patel
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Joy Wu
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Martin Urschler
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Amy Fong
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Cameron Simmers
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Gregory P Tarr
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Stuart Barnard
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| | - Ben Wilson
- Department of Radiology, Dunedin Hospital, 201 Great King St, Dunedin Central, Dunedin, Otago 9016, New Zealand (S.F., A.F., C.S., B.W.); Eastern Health, Melbourne, Victoria, Australia (D.A.); Auckland District Health Board, Auckland, New Zealand (J.S.K., J.Y., E.K., M.H.); Waitemata District Health Board, Auckland, New Zealand (C.K.J.); Waikato District Health Board, Hamilton, New Zealand (S.P.G.); The University of Auckland Faculty of Medical and Health Sciences, Auckland, Auckland, New Zealand (A.L.); University of Otago Medical School, Dunedin, Otago, New Zealand (A.P.); IBM Almaden Research Center, San Jose, Calif (J.W.); School of Computer Science, University of Auckland, Auckland, New Zealand (M.U.); Department of Radiology, Auckland City Hospital, Auckland, New Zealand (G.P.T.); and Department of Radiology, Middlemore Hospital, Auckland, New Zealand (S.B.)
| |
Collapse
|
125
|
Dahlblom V, Andersson I, Lång K, Tingberg A, Zackrisson S, Dustler M. Artificial Intelligence Detection of Missed Cancers at Digital Mammography That Were Detected at Digital Breast Tomosynthesis. Radiol Artif Intell 2021; 3:e200299. [PMID: 34870215 DOI: 10.1148/ryai.2021200299] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2020] [Revised: 07/12/2021] [Accepted: 08/09/2021] [Indexed: 11/11/2022]
Abstract
Purpose To investigate how an artificial intelligence (AI) system performs at digital mammography (DM) from a screening population with ground truth defined by digital breast tomosynthesis (DBT), and whether AI could detect breast cancers at DM that had originally only been detected at DBT. Materials and Methods In this secondary analysis of data from a prospective study, DM examinations from 14 768 women (mean age, 57 years), examined with both DM and DBT with independent double reading in the Malmӧ Breast Tomosynthesis Screening Trial (MBTST) (ClinicalTrials.gov: NCT01091545; data collection, 2010-2015), were analyzed with an AI system. Of 136 screening-detected cancers, 95 cancers were detected at DM and 41 cancers were detected only at DBT. The system identifies suspicious areas in the image, scored 1-100, and provides a risk score of 1 to 10 for the whole examination. A cancer was defined as AI detected if the cancer lesion was correctly localized and scored at least 62 (threshold determined by the AI system developers), therefore resulting in the highest examination risk score of 10. Data were analyzed with descriptive statistics, and detection performance was analyzed with receiver operating characteristics. Results The highest examination risk score was assigned to 10% (1493 of 14 786) of the examinations. With 90.8% specificity, the AI system detected 75% (71 of 95) of the DM-detected cancers and 44% (18 of 41) of cancers at DM that had originally been detected only at DBT. The majority were invasive cancers (17 of 18). Conclusion Almost half of the additional DBT-only screening-detected cancers in the MBTST were detected at DM with AI. AI did not reach double reading performance; however, if combined with double reading, AI has the potential to achieve a substantial portion of the benefit of DBT screening.Keywords: Computer-aided Diagnosis, Mammography, Breast, Diagnosis, Classification, Application DomainClinical trial registration no. NCT01091545© RSNA, 2021.
Collapse
Affiliation(s)
- Victor Dahlblom
- Diagnostic Radiology (V.D., I.A., K.L., S.Z., M.D.) and Medical Radiation Physics (A.T., M.D.), Department of Translational Medicine, Lund University, Malmö, Sweden; and Department of Medical Imaging and Physiology (V.D., S.Z.), Unilabs Breast Centre (I.A., K.L.), and Department of Radiation Physics (A.T.), Skåne University Hospital, Carl Bertil Laurells gata 9, 205 02 Malmö, Sweden
| | - Ingvar Andersson
- Diagnostic Radiology (V.D., I.A., K.L., S.Z., M.D.) and Medical Radiation Physics (A.T., M.D.), Department of Translational Medicine, Lund University, Malmö, Sweden; and Department of Medical Imaging and Physiology (V.D., S.Z.), Unilabs Breast Centre (I.A., K.L.), and Department of Radiation Physics (A.T.), Skåne University Hospital, Carl Bertil Laurells gata 9, 205 02 Malmö, Sweden
| | - Kristina Lång
- Diagnostic Radiology (V.D., I.A., K.L., S.Z., M.D.) and Medical Radiation Physics (A.T., M.D.), Department of Translational Medicine, Lund University, Malmö, Sweden; and Department of Medical Imaging and Physiology (V.D., S.Z.), Unilabs Breast Centre (I.A., K.L.), and Department of Radiation Physics (A.T.), Skåne University Hospital, Carl Bertil Laurells gata 9, 205 02 Malmö, Sweden
| | - Anders Tingberg
- Diagnostic Radiology (V.D., I.A., K.L., S.Z., M.D.) and Medical Radiation Physics (A.T., M.D.), Department of Translational Medicine, Lund University, Malmö, Sweden; and Department of Medical Imaging and Physiology (V.D., S.Z.), Unilabs Breast Centre (I.A., K.L.), and Department of Radiation Physics (A.T.), Skåne University Hospital, Carl Bertil Laurells gata 9, 205 02 Malmö, Sweden
| | - Sophia Zackrisson
- Diagnostic Radiology (V.D., I.A., K.L., S.Z., M.D.) and Medical Radiation Physics (A.T., M.D.), Department of Translational Medicine, Lund University, Malmö, Sweden; and Department of Medical Imaging and Physiology (V.D., S.Z.), Unilabs Breast Centre (I.A., K.L.), and Department of Radiation Physics (A.T.), Skåne University Hospital, Carl Bertil Laurells gata 9, 205 02 Malmö, Sweden
| | - Magnus Dustler
- Diagnostic Radiology (V.D., I.A., K.L., S.Z., M.D.) and Medical Radiation Physics (A.T., M.D.), Department of Translational Medicine, Lund University, Malmö, Sweden; and Department of Medical Imaging and Physiology (V.D., S.Z.), Unilabs Breast Centre (I.A., K.L.), and Department of Radiation Physics (A.T.), Skåne University Hospital, Carl Bertil Laurells gata 9, 205 02 Malmö, Sweden
| |
Collapse
|
126
|
Lee HJ, Nguyen AT, Ki SY, Lee JE, Do LN, Park MH, Lee JS, Kim HJ, Park I, Lim HS. Classification of MR-Detected Additional Lesions in Patients With Breast Cancer Using a Combination of Radiomics Analysis and Machine Learning. Front Oncol 2021; 11:744460. [PMID: 34926256 PMCID: PMC8679659 DOI: 10.3389/fonc.2021.744460] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2021] [Accepted: 11/08/2021] [Indexed: 01/02/2023] Open
Abstract
ObjectiveThis study was conducted in order to investigate the feasibility of using radiomics analysis (RA) with machine learning algorithms based on breast magnetic resonance (MR) images for discriminating malignant from benign MR-detected additional lesions in patients with primary breast cancer.Materials and MethodsOne hundred seventy-four MR-detected additional lesions (benign, n = 86; malignancy, n = 88) from 158 patients with ipsilateral primary breast cancer from a tertiary medical center were included in this retrospective study. The entire data were randomly split to training (80%) and independent test sets (20%). In addition, 25 patients (benign, n = 21; malignancy, n = 15) from another tertiary medical center were included for the external test. Radiomics features that were extracted from three regions-of-interest (ROIs; intratumor, peritumor, combined) using fat-saturated T1-weighted images obtained by subtracting pre- from postcontrast images (SUB) and T2-weighted image (T2) were utilized to train the support vector machine for the binary classification. A decision tree method was utilized to build a classifier model using clinical imaging interpretation (CII) features assessed by radiologists. Area under the receiver operating characteristic curve (AUROC), accuracy, sensitivity, and specificity were used to compare the diagnostic performance.ResultsThe RA models trained using radiomics features from the intratumor-ROI showed comparable performance to the CII model (accuracy, AUROC: 73.3%, 69.6% for the SUB RA model; 70.0%, 75.1% for the T2 RA model; 73.3%, 72.0% for the CII model). The diagnostic performance increased when the radiomics and CII features were combined to build a fusion model. The fusion model that combines the CII features and radiomics features from multiparametric MRI data demonstrated the highest performance with an accuracy of 86.7% and an AUROC of 91.1%. The external test showed a similar pattern where the fusion models demonstrated higher levels of performance compared with the RA- or CII-only models. The accuracy and AUROC of the SUB+T2 RA+CII model in the external test were 80.6% and 91.4%, respectively.ConclusionOur study demonstrated the feasibility of using RA with machine learning approach based on multiparametric MRI for quantitatively characterizing MR-detected additional lesions. The fusion model demonstrated an improved diagnostic performance over the models trained with either RA or CII alone.
Collapse
Affiliation(s)
- Hyo-jae Lee
- Department of Radiology, Chonnam National University Hospital, Gwangju, South Korea
| | - Anh-Tien Nguyen
- Department of Radiology, Chonnam National University Hospital, Gwangju, South Korea
| | - So Yeon Ki
- Department of Radiology, Chonnam National University Hwasun Hospital, Hwasun-gun, South Korea
| | - Jong Eun Lee
- Department of Radiology, Chonnam National University Hospital, Gwangju, South Korea
| | - Luu-Ngoc Do
- Department of Radiology, Chonnam National University, Gwangju, South Korea
| | - Min Ho Park
- Department of Radiology, Chonnam National University, Gwangju, South Korea
- Department of Surgery, Chonnam National University Hwasun Hospital, Hwasun-gun, South Korea
| | - Ji Shin Lee
- Department of Radiology, Chonnam National University, Gwangju, South Korea
- Department of Pathology, Chonnam National University Hwasun Hospital, Hwasun-gun, South Korea
| | - Hye Jung Kim
- Department of Radiology, School of Medicine, Kyungpook National University, Kyungpook National University Chilgok Hospital, Daegu, South Korea
| | - Ilwoo Park
- Department of Radiology, Chonnam National University Hospital, Gwangju, South Korea
- Department of Radiology, Chonnam National University, Gwangju, South Korea
- Department of Artificial Intelligence Convergence, Chonnam National University, Gwangju, South Korea
- *Correspondence: Ilwoo Park, ; Hyo Soon Lim,
| | - Hyo Soon Lim
- Department of Radiology, Chonnam National University Hwasun Hospital, Hwasun-gun, South Korea
- Department of Radiology, Chonnam National University, Gwangju, South Korea
- *Correspondence: Ilwoo Park, ; Hyo Soon Lim,
| |
Collapse
|
127
|
Harvey HB, Gowda V. Regulatory Issues and Challenges to Artificial Intelligence Adoption. Radiol Clin North Am 2021; 59:1075-1083. [PMID: 34689875 DOI: 10.1016/j.rcl.2021.07.007] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Abstract
Artificial intelligence technology promises to redefine the practice of radiology. However, it exists in a nascent phase and remains largely untested in the clinical space. This nature is both a cause and consequence of the uncertain legal-regulatory environment it enters. This discussion aims to shed light on these challenges, tracing the various pathways toward approval by the US Food and Drug Administration, the future of government oversight, privacy issues, ethical dilemmas, and practical considerations related to implementation in radiologist practice.
Collapse
Affiliation(s)
- Harlan Benjamin Harvey
- Radiology, Massachusetts General Hospital, Harvard Medical School, 175 Cambridge Street, Suite 200, Boston, MA 02114, USA.
| | - Vrushab Gowda
- Harvard Law School, 1563 Massachusetts Avenue, Cambridge, MA 02138, USA
| |
Collapse
|
128
|
Artificial intelligence for the real world of breast screening. Eur J Radiol 2021; 144:109661. [PMID: 34598013 DOI: 10.1016/j.ejrad.2021.109661] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2021] [Revised: 03/08/2021] [Accepted: 03/15/2021] [Indexed: 11/21/2022]
Abstract
Breast cancer screening with mammography reduces mortality in the women who attend by detecting high risk cancer early. It is far from perfect with variations in both sensitivity for the detection of cancer and very wide variations in specificity, leading to unnecessary recalls and biopsies. Over the last 12 months several papers have reported on AI algorithms that perform as well as human readers on large well curated population data sets. The nature of the test sets, the way the gold standard has been calculated, the definition of a positive call, and the statistics used all influence the results. Historically retrospective studies have not predicted the real-life performance of radiologist plus machine. So, it is important to perform prospective studies before introducing Artificial intelligence into real world breast screening.
Collapse
|
129
|
Yan Y, Schaffter T, Bergquist T, Yu T, Prosser J, Aydin Z, Jabeer A, Brugere I, Gao J, Chen G, Causey J, Yao Y, Bryson K, Long DR, Jarvik JG, Lee CI, Wilcox A, Guinney J, Mooney S. A Continuously Benchmarked and Crowdsourced Challenge for Rapid Development and Evaluation of Models to Predict COVID-19 Diagnosis and Hospitalization. JAMA Netw Open 2021; 4:e2124946. [PMID: 34633425 PMCID: PMC8506231 DOI: 10.1001/jamanetworkopen.2021.24946] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/21/2021] [Accepted: 07/08/2021] [Indexed: 01/28/2023] Open
Abstract
Importance Machine learning could be used to predict the likelihood of diagnosis and severity of illness. Lack of COVID-19 patient data has hindered the data science community in developing models to aid in the response to the pandemic. Objectives To describe the rapid development and evaluation of clinical algorithms to predict COVID-19 diagnosis and hospitalization using patient data by citizen scientists, provide an unbiased assessment of model performance, and benchmark model performance on subgroups. Design, Setting, and Participants This diagnostic and prognostic study operated a continuous, crowdsourced challenge using a model-to-data approach to securely enable the use of regularly updated COVID-19 patient data from the University of Washington by participants from May 6 to December 23, 2020. A postchallenge analysis was conducted from December 24, 2020, to April 7, 2021, to assess the generalizability of models on the cumulative data set as well as subgroups stratified by age, sex, race, and time of COVID-19 test. By December 23, 2020, this challenge engaged 482 participants from 90 teams and 7 countries. Main Outcomes and Measures Machine learning algorithms used patient data and output a score that represented the probability of patients receiving a positive COVID-19 test result or being hospitalized within 21 days after receiving a positive COVID-19 test result. Algorithms were evaluated using area under the receiver operating characteristic curve (AUROC) and area under the precision recall curve (AUPRC) scores. Ensemble models aggregating models from the top challenge teams were developed and evaluated. Results In the analysis using the cumulative data set, the best performance for COVID-19 diagnosis prediction was an AUROC of 0.776 (95% CI, 0.775-0.777) and an AUPRC of 0.297, and for hospitalization prediction, an AUROC of 0.796 (95% CI, 0.794-0.798) and an AUPRC of 0.188. Analysis on top models submitting to the challenge showed consistently better model performance on the female group than the male group. Among all age groups, the best performance was obtained for the 25- to 49-year age group, and the worst performance was obtained for the group aged 17 years or younger. Conclusions and Relevance In this diagnostic and prognostic study, models submitted by citizen scientists achieved high performance for the prediction of COVID-19 testing and hospitalization outcomes. Evaluation of challenge models on demographic subgroups and prospective data revealed performance discrepancies, providing insights into the potential bias and limitations in the models.
Collapse
Affiliation(s)
- Yao Yan
- Sage Bionetworks, Seattle, Washington
- Molecular Engineering and Sciences Institute, University of Washington, Seattle
| | | | - Timothy Bergquist
- Sage Bionetworks, Seattle, Washington
- Department of Biomedical Informatics and Medical Education, University of Washington, Seattle
| | - Thomas Yu
- Sage Bionetworks, Seattle, Washington
| | - Justin Prosser
- Institute of Translational Health Sciences, University of Washington, Seattle
| | - Zafer Aydin
- Department of Computer Engineering, Faculty of Engineering, Abdullah Gul University, Kayseri, Turkey
| | - Amhar Jabeer
- Department of Computer Engineering, Faculty of Engineering, Abdullah Gul University, Kayseri, Turkey
| | - Ivan Brugere
- Department of Computer Science, University of Illinois at Chicago, Chicago
| | - Jifan Gao
- Department of Biostatistics and Medical Informatics, University of Wisconsin–Madison, Madison
| | - Guanhua Chen
- Department of Biostatistics and Medical Informatics, University of Wisconsin–Madison, Madison
| | - Jason Causey
- Computer Science Department, College of Engineering and Computer Science, Arkansas State University, Jonesboro
- Arkansas AI-Campus, Center for No-Boundary Thinking, Arkansas State University, Jonesboro
| | - Yuxin Yao
- Department of Computer Science, University College London, London, United Kingdom
| | - Kevin Bryson
- Department of Computer Science, University College London, London, United Kingdom
| | - Dustin R. Long
- Division of Critical Care Medicine, Department of Anesthesiology and Pain Medicine, University of Washington, Seattle
| | - Jeffrey G. Jarvik
- The University of Washington Clinical Learning, Evidence And Research Center for Musculoskeletal Disorders, Seattle
- Department of Radiology, University of Washington School of Medicine, Seattle
| | - Christoph I. Lee
- Department of Radiology, University of Washington School of Medicine, Seattle
| | - Adam Wilcox
- Department of Biomedical Informatics and Medical Education, University of Washington, Seattle
| | | | - Sean Mooney
- Department of Biomedical Informatics and Medical Education, University of Washington, Seattle
| |
Collapse
|
130
|
Bhandari A, Purchuri SN, Sharma C, Ibrahim M, Prior M. Knowledge and attitudes towards artificial intelligence in imaging: a look at the quantitative survey literature. Clin Imaging 2021; 80:413-419. [PMID: 34537484 DOI: 10.1016/j.clinimag.2021.08.004] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/05/2021] [Revised: 07/31/2021] [Accepted: 08/05/2021] [Indexed: 10/20/2022]
Abstract
RATIONALE AND OBJECTIVES There exists many single sample perspectives on artificial intelligence (AI). The aim of this review was to collate the current data on attitudes/knowledge towards AI in three unique populations: medical students, clinicians and patients. MATERIALS AND METHODS A literature search was performed on PubMed, Scopus and Web of Science pertaining to survey data on AI in radiology. Quality assessment was performed by an adapted version of the assessment tool from the National Heart, Lung and Blood Institute for Observational Studies. RESULTS Fourteen studies were found on attitudes/knowledge towards AI in radiology. Four studies examined medical students, seven on clinicians and three on patient populations. Deficiencies in the literature mainly related to sampling bias. Students had anxiety relating to future job prospects. Clinicians were optimistic and viewed AI as an aid to the diagnosis and wanted to further their knowledge. Patients were concerned about the lack of human interaction and accountability during error. CONCLUSION Attitudes and knowledge regarding AI in radiology remains a topic that needs to be researched further and education given pertaining to its use in a clinical setting.
Collapse
Affiliation(s)
- Abhishta Bhandari
- Townsville University Hospital, 100 Angus Smith Drive, Douglas, Townsville, QLD, Australia.
| | | | - Chinmay Sharma
- Townsville University Hospital, 100 Angus Smith Drive, Douglas, Townsville, QLD, Australia
| | - Muhammad Ibrahim
- Townsville University Hospital, 100 Angus Smith Drive, Douglas, Townsville, QLD, Australia
| | - Marita Prior
- Royal Brisbane and Women's Hospital, Herston, Queensland, Australia
| |
Collapse
|
131
|
Bitencourt A, Daimiel Naranjo I, Lo Gullo R, Rossi Saccarelli C, Pinker K. AI-enhanced breast imaging: Where are we and where are we heading? Eur J Radiol 2021; 142:109882. [PMID: 34392105 PMCID: PMC8387447 DOI: 10.1016/j.ejrad.2021.109882] [Citation(s) in RCA: 35] [Impact Index Per Article: 11.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2021] [Revised: 07/15/2021] [Accepted: 07/26/2021] [Indexed: 12/22/2022]
Abstract
Significant advances in imaging analysis and the development of high-throughput methods that can extract and correlate multiple imaging parameters with different clinical outcomes have led to a new direction in medical research. Radiomics and artificial intelligence (AI) studies are rapidly evolving and have many potential applications in breast imaging, such as breast cancer risk prediction, lesion detection and classification, radiogenomics, and prediction of treatment response and clinical outcomes. AI has been applied to different breast imaging modalities, including mammography, ultrasound, and magnetic resonance imaging, in different clinical scenarios. The application of AI tools in breast imaging has an unprecedented opportunity to better derive clinical value from imaging data and reshape the way we care for our patients. The aim of this study is to review the current knowledge and future applications of AI-enhanced breast imaging in clinical practice.
Collapse
Affiliation(s)
- Almir Bitencourt
- Department of Imaging, A.C.Camargo Cancer Center, Sao Paulo, SP, Brazil; Dasa, Sao Paulo, SP, Brazil
| | - Isaac Daimiel Naranjo
- Department of Radiology, Breast Imaging Service, Guy's and St. Thomas' NHS Trust, Great Maze Pond, London, UK
| | - Roberto Lo Gullo
- Department of Radiology, Memorial Sloan Kettering Cancer Center, New York, NY, USA
| | | | - Katja Pinker
- Department of Radiology, Memorial Sloan Kettering Cancer Center, New York, NY, USA.
| |
Collapse
|
132
|
Freeman K, Geppert J, Stinton C, Todkill D, Johnson S, Clarke A, Taylor-Phillips S. Use of artificial intelligence for image analysis in breast cancer screening programmes: systematic review of test accuracy. BMJ 2021; 374:n1872. [PMID: 34470740 PMCID: PMC8409323 DOI: 10.1136/bmj.n1872] [Citation(s) in RCA: 93] [Impact Index Per Article: 31.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 01/28/2023]
Abstract
OBJECTIVE To examine the accuracy of artificial intelligence (AI) for the detection of breast cancer in mammography screening practice. DESIGN Systematic review of test accuracy studies. DATA SOURCES Medline, Embase, Web of Science, and Cochrane Database of Systematic Reviews from 1 January 2010 to 17 May 2021. ELIGIBILITY CRITERIA Studies reporting test accuracy of AI algorithms, alone or in combination with radiologists, to detect cancer in women's digital mammograms in screening practice, or in test sets. Reference standard was biopsy with histology or follow-up (for screen negative women). Outcomes included test accuracy and cancer type detected. STUDY SELECTION AND SYNTHESIS Two reviewers independently assessed articles for inclusion and assessed the methodological quality of included studies using the QUality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) tool. A single reviewer extracted data, which were checked by a second reviewer. Narrative data synthesis was performed. RESULTS Twelve studies totalling 131 822 screened women were included. No prospective studies measuring test accuracy of AI in screening practice were found. Studies were of poor methodological quality. Three retrospective studies compared AI systems with the clinical decisions of the original radiologist, including 79 910 women, of whom 1878 had screen detected cancer or interval cancer within 12 months of screening. Thirty four (94%) of 36 AI systems evaluated in these studies were less accurate than a single radiologist, and all were less accurate than consensus of two or more radiologists. Five smaller studies (1086 women, 520 cancers) at high risk of bias and low generalisability to the clinical context reported that all five evaluated AI systems (as standalone to replace radiologist or as a reader aid) were more accurate than a single radiologist reading a test set in the laboratory. In three studies, AI used for triage screened out 53%, 45%, and 50% of women at low risk but also 10%, 4%, and 0% of cancers detected by radiologists. CONCLUSIONS Current evidence for AI does not yet allow judgement of its accuracy in breast cancer screening programmes, and it is unclear where on the clinical pathway AI might be of most benefit. AI systems are not sufficiently specific to replace radiologist double reading in screening programmes. Promising results in smaller studies are not replicated in larger studies. Prospective studies are required to measure the effect of AI in clinical practice. Such studies will require clear stopping rules to ensure that AI does not reduce programme specificity. STUDY REGISTRATION Protocol registered as PROSPERO CRD42020213590.
Collapse
Affiliation(s)
- Karoline Freeman
- Division of Health Sciences, University of Warwick, Coventry, UK
| | - Julia Geppert
- Division of Health Sciences, University of Warwick, Coventry, UK
| | - Chris Stinton
- Division of Health Sciences, University of Warwick, Coventry, UK
| | - Daniel Todkill
- Division of Health Sciences, University of Warwick, Coventry, UK
| | - Samantha Johnson
- Division of Health Sciences, University of Warwick, Coventry, UK
| | - Aileen Clarke
- Division of Health Sciences, University of Warwick, Coventry, UK
| | | |
Collapse
|
133
|
The Fermi-Dirac distribution provides a calibrated probabilistic output for binary classifiers. Proc Natl Acad Sci U S A 2021; 118:2100761118. [PMID: 34413191 PMCID: PMC8403970 DOI: 10.1073/pnas.2100761118] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/18/2022] Open
Abstract
While it would be desirable that the output of binary classification algorithms be the probability that the classification is correct, most algorithms do not provide a method to calculate such a probability. We propose a probabilistic output for binary classifiers based on an unexpected mapping of the probability of correct classification to the probability of occupation of a fermion in a quantum system, known as the Fermi–Dirac distribution. This mapping allows us to compute the optimal threshold to separate predicted classes and to calculate statistical parameters necessary to estimate confidence intervals of performance metrics. Using this mapping we propose an ensemble learning algorithm. In short, the Fermi–Dirac distribution provides a calibrated probabilistic output for binary classification. Binary classification is one of the central problems in machine-learning research and, as such, investigations of its general statistical properties are of interest. We studied the ranking statistics of items in binary classification problems and observed that there is a formal and surprising relationship between the probability of a sample belonging to one of the two classes and the Fermi–Dirac distribution determining the probability that a fermion occupies a given single-particle quantum state in a physical system of noninteracting fermions. Using this equivalence, it is possible to compute a calibrated probabilistic output for binary classifiers. We show that the area under the receiver operating characteristics curve (AUC) in a classification problem is related to the temperature of an equivalent physical system. In a similar manner, the optimal decision threshold between the two classes is associated with the chemical potential of an equivalent physical system. Using our framework, we also derive a closed-form expression to calculate the variance for the AUC of a classifier. Finally, we introduce FiDEL (Fermi–Dirac-based ensemble learning), an ensemble learning algorithm that uses the calibrated nature of the classifier’s output probability to combine possibly very different classifiers.
Collapse
|
134
|
Kerschke L, Weigel S, Rodriguez-Ruiz A, Karssemeijer N, Heindel W. Using deep learning to assist readers during the arbitration process: a lesion-based retrospective evaluation of breast cancer screening performance. Eur Radiol 2021; 32:842-852. [PMID: 34383147 PMCID: PMC8794989 DOI: 10.1007/s00330-021-08217-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2021] [Revised: 06/04/2021] [Accepted: 07/13/2021] [Indexed: 11/30/2022]
Abstract
Objectives To evaluate if artificial intelligence (AI) can discriminate recalled benign from recalled malignant mammographic screening abnormalities to improve screening performance. Methods A total of 2257 full-field digital mammography screening examinations, obtained 2011–2013, of women aged 50–69 years which were recalled for further assessment of 295 malignant out of 305 truly malignant lesions and 2289 benign lesions after independent double-reading with arbitration, were included in this retrospective study. A deep learning AI system was used to obtain a score (0–95) for each recalled lesion, representing the likelihood of breast cancer. The sensitivity on the lesion level and the proportion of women without false-positive ratings (non-FPR) resulting under AI were estimated as a function of the classification cutoff and compared to that of human readers. Results Using a cutoff of 1, AI decreased the proportion of women with false-positives from 89.9 to 62.0%, non-FPR 11.1% vs. 38.0% (difference 26.9%, 95% confidence interval 25.1–28.8%; p < .001), preventing 30.1% of reader-induced false-positive recalls, while reducing sensitivity from 96.7 to 91.1% (5.6%, 3.1–8.0%) as compared to human reading. The positive predictive value of recall (PPV-1) increased from 12.8 to 16.5% (3.7%, 3.5–4.0%). In women with mass-related lesions (n = 900), the non-FPR was 14.2% for humans vs. 36.7% for AI (22.4%, 19.8–25.3%) at a sensitivity of 98.5% vs. 97.1% (1.5%, 0–3.5%). Conclusion The application of AI during consensus conference might especially help readers to reduce false-positive recalls of masses at the expense of a small sensitivity reduction. Prospective studies are needed to further evaluate the screening benefit of AI in practice. Key Points • Integrating the use of artificial intelligence in the arbitration process reduces benign recalls and increases the positive predictive value of recall at the expense of some sensitivity loss. • Application of the artificial intelligence system to aid the decision to recall a woman seems particularly beneficial for masses, where the system reaches comparable sensitivity to that of the readers, but with considerably reduced false-positives. • About one-fourth of all recalled malignant lesions are not automatically marked by the system such that their evaluation (AI score) must be retrieved manually by the reader. A thorough reading of screening mammograms by readers to identify suspicious lesions therefore remains mandatory.
Collapse
Affiliation(s)
- Laura Kerschke
- Institute of Biostatistics and Clinical Research, IBKF, University of Muenster, Schmeddingstrasse 56, 48149, Muenster, Germany.
| | - Stefanie Weigel
- Clinic for Radiology and Reference Center for Mammography Muenster, University of Muenster and University Hospital Muenster, Albert-Schweitzer-Campus 1, 48149, Muenster, Germany
| | | | - Nico Karssemeijer
- ScreenPoint Medical BV, Toernooiveld 300, 6525, EC, Nijmegen, The Netherlands
- Department of Radiology and Nuclear Medicine, Radboud University Medical Center, Geert Grooteplein Zuid 10, 6525, Nijmegen, GA, The Netherlands
| | - Walter Heindel
- Clinic for Radiology and Reference Center for Mammography Muenster, University of Muenster and University Hospital Muenster, Albert-Schweitzer-Campus 1, 48149, Muenster, Germany
| |
Collapse
|
135
|
Buda M, Saha A, Walsh R, Ghate S, Li N, Święcicki A, Lo JY, Mazurowski MA. A Data Set and Deep Learning Algorithm for the Detection of Masses and Architectural Distortions in Digital Breast Tomosynthesis Images. JAMA Netw Open 2021; 4:e2119100. [PMID: 34398205 PMCID: PMC8369362 DOI: 10.1001/jamanetworkopen.2021.19100] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/15/2022] Open
Abstract
IMPORTANCE Breast cancer screening is among the most common radiological tasks, with more than 39 million examinations performed each year. While it has been among the most studied medical imaging applications of artificial intelligence, the development and evaluation of algorithms are hindered by the lack of well-annotated, large-scale publicly available data sets. OBJECTIVES To curate, annotate, and make publicly available a large-scale data set of digital breast tomosynthesis (DBT) images to facilitate the development and evaluation of artificial intelligence algorithms for breast cancer screening; to develop a baseline deep learning model for breast cancer detection; and to test this model using the data set to serve as a baseline for future research. DESIGN, SETTING, AND PARTICIPANTS In this diagnostic study, 16 802 DBT examinations with at least 1 reconstruction view available, performed between August 26, 2014, and January 29, 2018, were obtained from Duke Health System and analyzed. From the initial cohort, examinations were divided into 4 groups and split into training and test sets for the development and evaluation of a deep learning model. Images with foreign objects or spot compression views were excluded. Data analysis was conducted from January 2018 to October 2020. EXPOSURES Screening DBT. MAIN OUTCOMES AND MEASURES The detection algorithm was evaluated with breast-based free-response receiver operating characteristic curve and sensitivity at 2 false positives per volume. RESULTS The curated data set contained 22 032 reconstructed DBT volumes that belonged to 5610 studies from 5060 patients with a mean (SD) age of 55 (11) years and 5059 (100.0%) women. This included 4 groups of studies: (1) 5129 (91.4%) normal studies; (2) 280 (5.0%) actionable studies, for which where additional imaging was needed but no biopsy was performed; (3) 112 (2.0%) benign biopsied studies; and (4) 89 studies (1.6%) with cancer. Our data set included masses and architectural distortions that were annotated by 2 experienced radiologists. Our deep learning model reached breast-based sensitivity of 65% (39 of 60; 95% CI, 56%-74%) at 2 false positives per DBT volume on a test set of 460 examinations from 418 patients. CONCLUSIONS AND RELEVANCE The large, diverse, and curated data set presented in this study could facilitate the development and evaluation of artificial intelligence algorithms for breast cancer screening by providing data for training as well as a common set of cases for model validation. The performance of the model developed in this study showed that the task remains challenging; its performance could serve as a baseline for future model development.
Collapse
Affiliation(s)
- Mateusz Buda
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Ashirbani Saha
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Ruth Walsh
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Sujata Ghate
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Nianyi Li
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Albert Święcicki
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
| | - Joseph Y. Lo
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
- Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina
| | - Maciej A. Mazurowski
- Department of Radiology, Duke University Medical Center, Durham, North Carolina
- Department of Electrical and Computer Engineering, Duke University, Durham, North Carolina
- Department of Computer Science, Duke University, Durham, North Carolina
- Department of Biostatistics and Bioinformatics, Duke University Medical Center, Durham, North Carolina
| |
Collapse
|
136
|
Kerner J, Dogan A, von Recum H. Machine learning and big data provide crucial insight for future biomaterials discovery and research. Acta Biomater 2021; 130:54-65. [PMID: 34087445 DOI: 10.1016/j.actbio.2021.05.053] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2020] [Revised: 05/24/2021] [Accepted: 05/25/2021] [Indexed: 02/06/2023]
Abstract
Machine learning have been widely adopted in a variety of fields including engineering, science, and medicine revolutionizing how data is collected, used, and stored. Their implementation has led to a drastic increase in the number of computational models for the prediction of various numerical, categorical, or association events given input variables. We aim to examine recent advances in the use of machine learning when applied to the biomaterial field. Specifically, quantitative structure properties relationships offer the unique ability to correlate microscale molecular descriptors to larger macroscale material properties. These new models can be broken down further into four categories: regression, classification, association, and clustering. We examine recent approaches and new uses of machine learning in the three major categories of biomaterials: metals, polymers, and ceramics for rapid property prediction and trend identification. While current research is promising, limitations in the form of lack of standardized reporting and available databases complicates the implementation of described models. Herein, we hope to provide a snapshot of the current state of the field and a beginner's guide to navigating the intersection of biomaterials research and machine learning. STATEMENT OF SIGNIFICANCE: Machine learning and its methods have found a variety of uses beyond the field of computer science but have largely been neglected by those in realm of biomaterials. Through the use of more computational methods, biomaterials development can be expediated while reducing the need for standard trial and error methods. Within, we introduce four basic models that readers can potentially apply to their current research as well as current applications within the field. Furthermore, we hope that this article may act as a "call to action" for readers to realize and address the current lack of implementation within the biomaterials field.
Collapse
Affiliation(s)
- Jacob Kerner
- Case Western Reserve University; 10900 Euclid Ave., Cleveland Ohio 44106.
| | - Alan Dogan
- Case Western Reserve University; 10900 Euclid Ave., Cleveland Ohio 44106.
| | - Horst von Recum
- Case Western Reserve University; 10900 Euclid Ave., Cleveland Ohio 44106.
| |
Collapse
|
137
|
Yala A, Mikhael PG, Strand F, Lin G, Smith K, Wan YL, Lamb L, Hughes K, Lehman C, Barzilay R. Toward robust mammography-based models for breast cancer risk. Sci Transl Med 2021; 13:13/578/eaba4373. [PMID: 33504648 DOI: 10.1126/scitranslmed.aba4373] [Citation(s) in RCA: 77] [Impact Index Per Article: 25.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2020] [Revised: 07/24/2020] [Accepted: 12/21/2020] [Indexed: 12/14/2022]
Abstract
Improved breast cancer risk models enable targeted screening strategies that achieve earlier detection and less screening harm than existing guidelines. To bring deep learning risk models to clinical practice, we need to further refine their accuracy, validate them across diverse populations, and demonstrate their potential to improve clinical workflows. We developed Mirai, a mammography-based deep learning model designed to predict risk at multiple timepoints, leverage potentially missing risk factor information, and produce predictions that are consistent across mammography machines. Mirai was trained on a large dataset from Massachusetts General Hospital (MGH) in the United States and tested on held-out test sets from MGH, Karolinska University Hospital in Sweden, and Chang Gung Memorial Hospital (CGMH) in Taiwan, obtaining C-indices of 0.76 (95% confidence interval, 0.74 to 0.80), 0.81 (0.79 to 0.82), and 0.79 (0.79 to 0.83), respectively. Mirai obtained significantly higher 5-year ROC AUCs than the Tyrer-Cuzick model ( P < 0.001) and prior deep learning models Hybrid DL ( P < 0.001) and Image-Only DL ( P < 0.001), trained on the same dataset. Mirai more accurately identified high-risk patients than prior methods across all datasets. On the MGH test set, 41.5% (34.4 to 48.5) of patients who would develop cancer within 5 years were identified as high risk, compared with 36.1% (29.1 to 42.9) by Hybrid DL ( P = 0.02) and 22.9% (15.9 to 29.6) by the Tyrer-Cuzick model ( P < 0.001).
Collapse
Affiliation(s)
- Adam Yala
- Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, MA 02139, USA. .,Jameel Clinic, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Peter G Mikhael
- Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, MA 02139, USA.,Jameel Clinic, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Fredrik Strand
- Breast Radiology Unit, Department of Imaging and Physiology, Karolinska University Hospital, 17164 Solna, Sweden.,Department of Oncology-Pathology, Karolinska Institute, 17164 Solna, Sweden
| | - Gigin Lin
- Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou, Taoyuan 333, Taiwan
| | - Kevin Smith
- School of Electrical Engineering and Computer, KTH Royal Institute of Technology, 10044 Stockholm, Sweden.,Science for Life Laboratory, 17165 Solna, Sweden
| | - Yung-Liang Wan
- Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou, Taoyuan 333, Taiwan
| | - Leslie Lamb
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
| | - Kevin Hughes
- Division of Surgical Oncology, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
| | - Constance Lehman
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
| | - Regina Barzilay
- Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, MA 02139, USA.,Jameel Clinic, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| |
Collapse
|
138
|
Lee JI, Kim DH, Yoo HJ, Choi HG, Lee YS. Comparison of the Predicting Performance for Fate of Medial Meniscus Posterior Root Tear Based on Treatment Strategies: A Comparison between Logistic Regression, Gradient Boosting, and CNN Algorithms. Diagnostics (Basel) 2021; 11:diagnostics11071225. [PMID: 34359308 PMCID: PMC8304966 DOI: 10.3390/diagnostics11071225] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2021] [Accepted: 07/05/2021] [Indexed: 01/17/2023] Open
Abstract
This study aimed to validate the accuracy and prediction performance of machine learning (ML), deep learning (DL), and logistic regression methods in the treatment of medial meniscus posterior root tears (MMPRT). From July 2003 to May 2018, 640 patients diagnosed with MMPRT were included. First, the affecting factors for the surgery were evaluated using statistical analysis. Second, AI technology was introduced using X-ray and MRI. Finally, the accuracy and prediction performance were compared between ML&DL and logistic regression methods. Affecting factors of the logistic regression method corresponded well with the feature importance of the six top-ranked factors in the ML&DL method. There was no significant difference when comparing the accuracy, F1-score, and error rate between ML&DL and logistic regression methods (accuracy = 0.89 and 0.91, F1 score = 0.89 and 0.90, error rate = 0.11 and 0.09; p = 0.114, 0.422, and 0.119, respectively). The area under the curve (AUC) values showed excellent test quality for both ML&DL and logistic regression methods (AUC = 0.97 and 0.94, respectively) in the evaluation of prediction performance (p = 0.289). The affecting factors of the logistic regression method and the influence of the ML&DL method were not significantly different. The accuracy and performance of the ML&DL method in predicting the fate of MMPRT were comparable to those of the logistic regression method. Therefore, this ML&DL algorithm could potentially predict the outcome of the MMRPT in various fields and situations. Furthermore, our method could be efficiently implemented in current clinical practice.
Collapse
|
139
|
Frazer HM, Qin AK, Pan H, Brotchie P. Evaluation of deep learning-based artificial intelligence techniques for breast cancer detection on mammograms: Results from a retrospective study using a BreastScreen Victoria dataset. J Med Imaging Radiat Oncol 2021; 65:529-537. [PMID: 34212526 PMCID: PMC8456839 DOI: 10.1111/1754-9485.13278] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2021] [Accepted: 06/15/2021] [Indexed: 11/28/2022]
Abstract
Introduction This study aims to evaluate deep learning (DL)‐based artificial intelligence (AI) techniques for detecting the presence of breast cancer on a digital mammogram image. Methods We evaluated several DL‐based AI techniques that employ different approaches and backbone DL models and tested the effect on performance of using different data‐processing strategies on a set of digital mammographic images with annotations of pathologically proven breast cancer. Results Our evaluation uses the area under curve (AUC) and accuracy (ACC) for performance measurement. The best evaluation result, based on 349 test cases (930 test images), was an AUC of 0.8979 [95% confidence interval (CI) 0.873, 0.923] and ACC of 0.8178 [95% CI 0.785, 0.850]. This was achieved by an AI technique that utilises a certain family of DL models, namely ResNet, as its backbone, combines the global features extracted from the whole mammogram and the local features extracted from the automatically detected cancer and non‐cancer local regions in the whole image, and leverages background cropping and text removal, contrast adjustment and more training data. Conclusion DL‐based AI techniques have shown promising results in retrospective studies for many medical image analysis applications. Our study demonstrates a significant opportunity to boost the performance of such techniques applied to breast cancer detection by exploring different types of approaches, backbone DL models and data‐processing strategies. The promising results we have obtained suggest further development of AI reading services could transform breast cancer screening in the future.
Collapse
Affiliation(s)
- Helen Ml Frazer
- St Vincent's BreastScreen, St Vincent's Hospital Melbourne, Melbourne, Victoria, Australia
| | - Alex K Qin
- Department of Computer Science and Software Engineering, Swinburne University of Technology, Melbourne, Victoria, Australia
| | - Hong Pan
- Department of Computer Science and Software Engineering, Swinburne University of Technology, Melbourne, Victoria, Australia
| | - Peter Brotchie
- Department of Medical Imaging, St Vincent's Hospital Melbourne, Melbourne, Victoria, Australia
| |
Collapse
|
140
|
Hickman SE, Baxter GC, Gilbert FJ. Adoption of artificial intelligence in breast imaging: evaluation, ethical constraints and limitations. Br J Cancer 2021; 125:15-22. [PMID: 33772149 PMCID: PMC8257639 DOI: 10.1038/s41416-021-01333-w] [Citation(s) in RCA: 25] [Impact Index Per Article: 8.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2020] [Revised: 02/15/2021] [Accepted: 02/24/2021] [Indexed: 02/07/2023] Open
Abstract
Retrospective studies have shown artificial intelligence (AI) algorithms can match as well as enhance radiologist's performance in breast screening. These tools can facilitate tasks not feasible by humans such as the automatic triage of patients and prediction of treatment outcomes. Breast imaging faces growing pressure with the exponential growth in imaging requests and a predicted reduced workforce to provide reports. Solutions to alleviate these pressures are being sought with an increasing interest in the adoption of AI to improve workflow efficiency as well as patient outcomes. Vast quantities of data are needed to test and monitor AI algorithms before and after their incorporation into healthcare systems. Availability of data is currently limited, although strategies are being devised to harness the data that already exists within healthcare institutions. Challenges that underpin the realisation of AI into everyday breast imaging cannot be underestimated and the provision of guidance from national agencies to tackle these challenges, taking into account views from a societal, industrial and healthcare prospective is essential. This review provides background on the evaluation and use of AI in breast imaging in addition to exploring key ethical, technical, legal and regulatory challenges that have been identified so far.
Collapse
Affiliation(s)
- Sarah E Hickman
- Department of Radiology, University of Cambridge School of Clinical Medicine, Cambridge, UK
| | - Gabrielle C Baxter
- Department of Radiology, University of Cambridge School of Clinical Medicine, Cambridge, UK
| | - Fiona J Gilbert
- Department of Radiology, University of Cambridge School of Clinical Medicine, Cambridge, UK.
- Department of Radiology, Addenbrookes Hospital, Cambridge University Hospitals NHS Foundation Trust, Cambridge, UK.
| |
Collapse
|
141
|
Parums DV. Editorial: Artificial Intelligence (AI) in Clinical Medicine and the 2020 CONSORT-AI Study Guidelines. Med Sci Monit 2021; 27:e933675. [PMID: 34176921 PMCID: PMC8252890 DOI: 10.12659/msm.933675] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2021] [Accepted: 06/21/2021] [Indexed: 12/21/2022] Open
Abstract
Artificial intelligence (AI) in clinical medicine includes physical robotics and devices and virtual AI and machine learning. Concerns have been raised regarding ethical issues for the use of AI in surgery, including guidance for surgical decisions, patient confidentiality, and the need for support from controlled clinical trials to use these methods so that clinical guidelines can be developed. The most common applications for virtual AI include disease diagnosis, health monitoring and digital patient consultations, clinical training, patient data management, drug development, and personalized medicine. In September 2020, the CONSORT-A1 extension was developed with 14 additional items that should be reported for AI studies that include clear descriptions of the AI intervention, skills required, study setting, inputs and outputs of the AI intervention, analysis of errors, and the human and AI interactions. This Editorial aims to present current applications and challenges of AI in clinical medicine and the importance of the new 2020 CONSORT-AI study guidelines.
Collapse
Affiliation(s)
- Dinah V Parums
- Science Editor, Medical Science Monitor, International Scientific Information, Inc., Mellville, NY, USA
| |
Collapse
|
142
|
Thrall JH, Fessell D, Pandharipande PV. Rethinking the Approach to Artificial Intelligence for Medical Image Analysis: The Case for Precision Diagnosis. J Am Coll Radiol 2021; 18:174-179. [PMID: 33413896 DOI: 10.1016/j.jacr.2020.07.010] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2020] [Accepted: 07/07/2020] [Indexed: 02/08/2023]
Abstract
To date, widely generalizable artificial intelligence (AI) programs for medical image analysis have not been demonstrated, including for mammography. Rather than pursuing a strategy of collecting ever-larger databases in the attempt to build generalizable programs, we suggest three possible avenues for exploring a precision medicine or precision imaging approach. First, it is now technologically feasible to collect hundreds of thousands of multi-institutional cases along with other patient data, allowing stratification of patients into subpopulations that have similar characteristics in the manner discussed by the National Research Council in its white paper on precision medicine. A family of AI programs could be developed across different examination types that are matched to specific patient subpopulations. Such stratification can help address bias, including racial or ethnic bias, by allowing unbiased data aggregation for creation of subpopulations. Second, for common examinations, larger institutions may be able to collect enough of their own data to train AI programs that reflect disease prevalence and variety in their respective unique patient subpopulations. Third, high- and low-probability subpopulations can be identified by application of AI programs, thereby allowing their triage off the radiology work list. This would reduce radiologists' workloads, providing more time for interpretation of the remaining examinations. For high-volume procedures, investigators should come together to define reference standards, collect data, and compare the merits of pursuing generalizability versus a precision medicine subpopulation-based strategy.
Collapse
Affiliation(s)
- James H Thrall
- Chair Emeritus, Department of Radiology, Massachusetts General Hospital, Boston, Massachusetts.
| | - David Fessell
- Associate Professor, Department of Radiology, University of Michigan, Ann Arbor, Michigan
| | - Pari V Pandharipande
- Director, MGH Institute for Technology Assessment; Associate Chair, Integrated Imaging & Imaging Sciences, MGH Radiology; Executive Director, Clinical Enterprise Integration, Mass General Brigham (MGB) Radiology, Boston, Massachusetts
| |
Collapse
|
143
|
Strand F, Patel BK, Allen B. A Call for Controlled Validation Data Sets: Promoting the Safe Introduction of Artificial Intelligence in Breast Imaging. J Am Coll Radiol 2021; 18:1564-1565. [PMID: 34147505 DOI: 10.1016/j.jacr.2021.06.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2021] [Revised: 05/25/2021] [Accepted: 06/02/2021] [Indexed: 10/21/2022]
Affiliation(s)
- Fredrik Strand
- Breast Radiology, Karolinska University Hospital, Solna, Sweden; Department of Oncology-Pathology, Karolinska Institute, Solna, Sweden.
| | - Bhavika K Patel
- Breast Imaging Panel Cochair for the ACR Data Science Institute; Chair of Radiology Research, Diagnostic Radiology, Mayo Clinic Arizona Phoenix, Arizona
| | - Bibb Allen
- Chief Medical Officer of the ACR Data Science Institute and is from Diagnostic Radiology, Grandview Medical Center, Birmingham, Alabama; Program Director, Diagnostic Radiology Residency Program, Brookwood Baptist Health, Birmingham, Alabama
| |
Collapse
|
144
|
Rudie JD, Duda J, Duong MT, Chen PH, Xie L, Kurtz R, Ware JB, Choi J, Mattay RR, Botzolakis EJ, Gee JC, Bryan RN, Cook TS, Mohan S, Nasrallah IM, Rauschecker AM. Brain MRI Deep Learning and Bayesian Inference System Augments Radiology Resident Performance. J Digit Imaging 2021; 34:1049-1058. [PMID: 34131794 DOI: 10.1007/s10278-021-00470-1] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2020] [Revised: 04/28/2021] [Accepted: 05/25/2021] [Indexed: 12/15/2022] Open
Abstract
Automated quantitative and probabilistic medical image analysis has the potential to improve the accuracy and efficiency of the radiology workflow. We sought to determine whether AI systems for brain MRI diagnosis could be used as a clinical decision support tool to augment radiologist performance. We utilized previously developed AI systems that combine convolutional neural networks and expert-derived Bayesian networks to distinguish among 50 diagnostic entities on multimodal brain MRIs. We tested whether these systems could augment radiologist performance through an interactive clinical decision support tool known as Adaptive Radiology Interpretation and Education System (ARIES) in 194 test cases. Four radiology residents and three academic neuroradiologists viewed half of the cases unassisted and half with the results of the AI system displayed on ARIES. Diagnostic accuracy of radiologists for top diagnosis (TDx) and top three differential diagnosis (T3DDx) was compared with and without ARIES. Radiology resident performance was significantly better with ARIES for both TDx (55% vs 30%; P < .001) and T3DDx (79% vs 52%; P = 0.002), with the largest improvement for rare diseases (39% increase for T3DDx; P < 0.001). There was no significant difference between attending performance with and without ARIES for TDx (72% vs 69%; P = 0.48) or T3DDx (86% vs 89%; P = 0.39). These findings suggest that a hybrid deep learning and Bayesian inference clinical decision support system has the potential to augment diagnostic accuracy of non-specialists to approach the level of subspecialists for a large array of diseases on brain MRI.
Collapse
Affiliation(s)
- Jeffrey D Rudie
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA. .,Department of Radiology & Biomedical Imaging, University of California, San Francisco, San Francisco, CA, USA.
| | - Jeffrey Duda
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - Michael Tran Duong
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - Po-Hao Chen
- Department of Radiology, Cleveland Clinic Imaging Institute, Cleveland, OH, USA
| | - Long Xie
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - Robert Kurtz
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - Jeffrey B Ware
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - Joshua Choi
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - Raghav R Mattay
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | | | - James C Gee
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - R Nick Bryan
- Department of Diagnostic Medicine, Dell Medical School, University of Texas, Austin, TX, USA
| | - Tessa S Cook
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - Suyash Mohan
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - Ilya M Nasrallah
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA
| | - Andreas M Rauschecker
- Department of Radiology, Perelman School of Medicine at the University of Pennsylvania, PA, Philadelphia, USA.,Department of Radiology & Biomedical Imaging, University of California, San Francisco, San Francisco, CA, USA
| |
Collapse
|
145
|
Tarca AL, Pataki BÁ, Romero R, Sirota M, Guan Y, Kutum R, Gomez-Lopez N, Done B, Bhatti G, Yu T, Andreoletti G, Chaiworapongsa T, Hassan SS, Hsu CD, Aghaeepour N, Stolovitzky G, Csabai I, Costello JC. Crowdsourcing assessment of maternal blood multi-omics for predicting gestational age and preterm birth. Cell Rep Med 2021; 2:100323. [PMID: 34195686 PMCID: PMC8233692 DOI: 10.1016/j.xcrm.2021.100323] [Citation(s) in RCA: 34] [Impact Index Per Article: 11.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2020] [Revised: 01/18/2021] [Accepted: 05/20/2021] [Indexed: 12/15/2022]
Abstract
Identification of pregnancies at risk of preterm birth (PTB), the leading cause of newborn deaths, remains challenging given the syndromic nature of the disease. We report a longitudinal multi-omics study coupled with a DREAM challenge to develop predictive models of PTB. The findings indicate that whole-blood gene expression predicts ultrasound-based gestational ages in normal and complicated pregnancies (r = 0.83) and, using data collected before 37 weeks of gestation, also predicts the delivery date in both normal pregnancies (r = 0.86) and those with spontaneous preterm birth (r = 0.75). Based on samples collected before 33 weeks in asymptomatic women, our analysis suggests that expression changes preceding preterm prelabor rupture of the membranes are consistent across time points and cohorts and involve leukocyte-mediated immunity. Models built from plasma proteomic data predict spontaneous preterm delivery with intact membranes with higher accuracy and earlier in pregnancy than transcriptomic models (AUROC = 0.76 versus AUROC = 0.6 at 27-33 weeks of gestation).
Collapse
Affiliation(s)
- Adi L. Tarca
- Perinatology Research Branch, Division of Obstetrics and Maternal-Fetal Medicine, Division of Intramural Research, Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, US Department of Health and Human Services, Detroit, MI 48201, USA
- Department of Obstetrics and Gynecology, Wayne State University School of Medicine, Detroit, MI 48201 USA
- Department of Computer Science, Wayne State University College of Engineering, Detroit, MI 48202, USA
| | - Bálint Ármin Pataki
- Department of Physics of Complex Systems, ELTE Eötvös Loránd University, Budapest, Hungary
| | - Roberto Romero
- Perinatology Research Branch, Division of Obstetrics and Maternal-Fetal Medicine, Division of Intramural Research, Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, US Department of Health and Human Services, Detroit, MI 48201, USA
- Department of Obstetrics and Gynecology, University of Michigan, Ann Arbor, MI 48109, USA
- Department of Epidemiology and Biostatistics, Michigan State University, East Lansing, MI 48824, USA
- Center for Molecular Medicine and Genetics, Wayne State University, Detroit, MI 48201, USA
- Detroit Medical Center, Detroit, MI 48201, USA
- Department of Obstetrics and Gynecology, Florida International University, Miami, FL 33199, USA
| | - Marina Sirota
- Department of Pediatrics, University of California, San Francisco, San Francisco, CA 94143, USA
- Bakar Computational Health Sciences Institute, University of California, San Francisco, San Francisco, CA 94158, USA
| | - Yuanfang Guan
- Department of Computational Medicine and Bioinformatics, University of Michigan, Ann Arbor, MI 48109, USA
| | - Rintu Kutum
- Informatics and Big Data Unit, CSIR-Institute of Genomics and Integrative Biology, New Delhi, India
| | - Nardhy Gomez-Lopez
- Perinatology Research Branch, Division of Obstetrics and Maternal-Fetal Medicine, Division of Intramural Research, Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, US Department of Health and Human Services, Detroit, MI 48201, USA
- Department of Obstetrics and Gynecology, Wayne State University School of Medicine, Detroit, MI 48201 USA
- Department of Biochemistry, Microbiology, and Immunology, Wayne State University School of Medicine, Detroit, MI 48201 USA
| | - Bogdan Done
- Perinatology Research Branch, Division of Obstetrics and Maternal-Fetal Medicine, Division of Intramural Research, Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, US Department of Health and Human Services, Detroit, MI 48201, USA
- Department of Obstetrics and Gynecology, Wayne State University School of Medicine, Detroit, MI 48201 USA
| | - Gaurav Bhatti
- Perinatology Research Branch, Division of Obstetrics and Maternal-Fetal Medicine, Division of Intramural Research, Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, US Department of Health and Human Services, Detroit, MI 48201, USA
- Department of Obstetrics and Gynecology, Wayne State University School of Medicine, Detroit, MI 48201 USA
| | | | - Gaia Andreoletti
- Department of Pediatrics, University of California, San Francisco, San Francisco, CA 94143, USA
- Bakar Computational Health Sciences Institute, University of California, San Francisco, San Francisco, CA 94158, USA
| | - Tinnakorn Chaiworapongsa
- Department of Obstetrics and Gynecology, Wayne State University School of Medicine, Detroit, MI 48201 USA
| | - The DREAM Preterm Birth Prediction Challenge Consortium
- Perinatology Research Branch, Division of Obstetrics and Maternal-Fetal Medicine, Division of Intramural Research, Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, US Department of Health and Human Services, Detroit, MI 48201, USA
- Department of Obstetrics and Gynecology, Wayne State University School of Medicine, Detroit, MI 48201 USA
- Department of Computer Science, Wayne State University College of Engineering, Detroit, MI 48202, USA
- Department of Physics of Complex Systems, ELTE Eötvös Loránd University, Budapest, Hungary
- Department of Obstetrics and Gynecology, University of Michigan, Ann Arbor, MI 48109, USA
- Department of Epidemiology and Biostatistics, Michigan State University, East Lansing, MI 48824, USA
- Center for Molecular Medicine and Genetics, Wayne State University, Detroit, MI 48201, USA
- Detroit Medical Center, Detroit, MI 48201, USA
- Department of Obstetrics and Gynecology, Florida International University, Miami, FL 33199, USA
- Department of Pediatrics, University of California, San Francisco, San Francisco, CA 94143, USA
- Bakar Computational Health Sciences Institute, University of California, San Francisco, San Francisco, CA 94158, USA
- Department of Computational Medicine and Bioinformatics, University of Michigan, Ann Arbor, MI 48109, USA
- Informatics and Big Data Unit, CSIR-Institute of Genomics and Integrative Biology, New Delhi, India
- Department of Biochemistry, Microbiology, and Immunology, Wayne State University School of Medicine, Detroit, MI 48201 USA
- Sage Bionetworks, Seattle, WA, USA
- Office of Women’s Health, Integrative Biosciences Center, Wayne State University, Detroit, MI 48202, USA
- Department of Physiology, Wayne State University School of Medicine, Detroit, MI 48201, USA
- Department of Anesthesiology, Perioperative, and Pain Medicine, Department of Pediatrics, and Department of Biomedical Data Sciences, Stanford University School of Medicine, Stanford, CA 94305, USA
- Department of Genetics and Genomic Sciences, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
- IBM T.J. Watson Research Center, Yorktown Heights, NY 10598, USA
- Department of Pharmacology, University of Colorado Anschutz Medical Campus, Aurora, CO 80045, USA
| | - Sonia S. Hassan
- Department of Obstetrics and Gynecology, Wayne State University School of Medicine, Detroit, MI 48201 USA
- Office of Women’s Health, Integrative Biosciences Center, Wayne State University, Detroit, MI 48202, USA
- Department of Physiology, Wayne State University School of Medicine, Detroit, MI 48201, USA
| | - Chaur-Dong Hsu
- Perinatology Research Branch, Division of Obstetrics and Maternal-Fetal Medicine, Division of Intramural Research, Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, US Department of Health and Human Services, Detroit, MI 48201, USA
- Department of Obstetrics and Gynecology, Wayne State University School of Medicine, Detroit, MI 48201 USA
- Department of Physiology, Wayne State University School of Medicine, Detroit, MI 48201, USA
| | - Nima Aghaeepour
- Department of Anesthesiology, Perioperative, and Pain Medicine, Department of Pediatrics, and Department of Biomedical Data Sciences, Stanford University School of Medicine, Stanford, CA 94305, USA
| | - Gustavo Stolovitzky
- Department of Genetics and Genomic Sciences, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
- IBM T.J. Watson Research Center, Yorktown Heights, NY 10598, USA
| | - Istvan Csabai
- Department of Physics of Complex Systems, ELTE Eötvös Loránd University, Budapest, Hungary
| | - James C. Costello
- Department of Pharmacology, University of Colorado Anschutz Medical Campus, Aurora, CO 80045, USA
| |
Collapse
|
146
|
Hendrix N, Hauber B, Lee CI, Bansal A, Veenstra DL. Artificial intelligence in breast cancer screening: primary care provider preferences. J Am Med Inform Assoc 2021; 28:1117-1124. [PMID: 33367670 PMCID: PMC8200265 DOI: 10.1093/jamia/ocaa292] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2020] [Revised: 10/05/2020] [Accepted: 11/10/2020] [Indexed: 11/15/2022] Open
Abstract
BACKGROUND Artificial intelligence (AI) is increasingly being proposed for use in medicine, including breast cancer screening (BCS). Little is known, however, about referring primary care providers' (PCPs') preferences for this technology. METHODS We identified the most important attributes of AI BCS for ordering PCPs using qualitative interviews: sensitivity, specificity, radiologist involvement, understandability of AI decision-making, supporting evidence, and diversity of training data. We invited US-based PCPs to participate in an internet-based experiment designed to force participants to trade off among the attributes of hypothetical AI BCS products. Responses were analyzed with random parameters logit and latent class models to assess how different attributes affect the choice to recommend AI-enhanced screening. RESULTS Ninety-one PCPs participated. Sensitivity was most important, and most PCPs viewed radiologist participation in mammography interpretation as important. Other important attributes were specificity, understandability of AI decision-making, and diversity of data. We identified 3 classes of respondents: "Sensitivity First" (41%) found sensitivity to be more than twice as important as other attributes; "Against AI Autonomy" (24%) wanted radiologists to confirm every image; "Uncertain Trade-Offs" (35%) viewed most attributes as having similar importance. A majority (76%) accepted the use of AI in a "triage" role that would allow it to filter out likely negatives without radiologist confirmation. CONCLUSIONS AND RELEVANCE Sensitivity was the most important attribute overall, but other key attributes should be addressed to produce clinically acceptable products. We also found that most PCPs accept the use of AI to make determinations about likely negative mammograms without radiologist confirmation.
Collapse
Affiliation(s)
- Nathaniel Hendrix
- The Comparative Health Outcomes, Policy & Economics (CHOICE) Institute, University of Washington School of Pharmacy, Seattle, Washington, USA
| | - Brett Hauber
- The Comparative Health Outcomes, Policy & Economics (CHOICE) Institute, University of Washington School of Pharmacy, Seattle, Washington, USA
- RTI Health Solutions, Research Triangle Park, North Carolina, USA
| | - Christoph I Lee
- Department of Radiology, University of Washington School of Medicine, Seattle, Washington, USA
- Department of Health Services, University of Washington School of Public Health, Seattle, Washington, USA
- Hutchinson Institute for Cancer Outcomes Research, Seattle, Washington, USA
| | - Aasthaa Bansal
- The Comparative Health Outcomes, Policy & Economics (CHOICE) Institute, University of Washington School of Pharmacy, Seattle, Washington, USA
| | - David L Veenstra
- The Comparative Health Outcomes, Policy & Economics (CHOICE) Institute, University of Washington School of Pharmacy, Seattle, Washington, USA
| |
Collapse
|
147
|
Born J, Beymer D, Rajan D, Coy A, Mukherjee VV, Manica M, Prasanna P, Ballah D, Guindy M, Shaham D, Shah PL, Karteris E, Robertus JL, Gabrani M, Rosen-Zvi M. On the role of artificial intelligence in medical imaging of COVID-19. PATTERNS (NEW YORK, N.Y.) 2021; 2:100269. [PMID: 33969323 PMCID: PMC8086827 DOI: 10.1016/j.patter.2021.100269] [Citation(s) in RCA: 27] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
Abstract
Although a plethora of research articles on AI methods on COVID-19 medical imaging are published, their clinical value remains unclear. We conducted the largest systematic review of the literature addressing the utility of AI in imaging for COVID-19 patient care. By keyword searches on PubMed and preprint servers throughout 2020, we identified 463 manuscripts and performed a systematic meta-analysis to assess their technical merit and clinical relevance. Our analysis evidences a significant disparity between clinical and AI communities, in the focus on both imaging modalities (AI experts neglected CT and ultrasound, favoring X-ray) and performed tasks (71.9% of AI papers centered on diagnosis). The vast majority of manuscripts were found to be deficient regarding potential use in clinical practice, but 2.7% (n = 12) publications were assigned a high maturity level and are summarized in greater detail. We provide an itemized discussion of the challenges in developing clinically relevant AI solutions with recommendations and remedies.
Collapse
Affiliation(s)
- Jannis Born
- IBM Research Europe, Zurich, Switzerland
- Department for Biosystems Science & Engineering, ETH Zurich, Zurich, Switzerland
| | | | | | - Adam Coy
- IBM Almaden Research Center, San Jose, CA, USA
- Vision Radiology, Dallas, TX, USA
| | | | | | - Prasanth Prasanna
- IBM Almaden Research Center, San Jose, CA, USA
- Department of Radiology and Imaging Sciences, University of Utah Health Sciences Center, Salt Lake City, UT, USA
| | - Deddeh Ballah
- IBM Almaden Research Center, San Jose, CA, USA
- Department of Radiology, Seton Medical Center, Daly City, CA, USA
| | - Michal Guindy
- Assuta Medical Centres Radiology, Tel-Aviv, Israel
- Ben-Gurion University Medical School, Be'er Sheva, Israel
| | - Dorith Shaham
- Department of Radiology, Hadassah-Hebrew University Medical Center, Faculty of Medicine, Hebrew University of Jerusalem, Jerusalem, Israel
| | - Pallav L. Shah
- Royal Brompton and Harefield Hospitals, Guy's and St Thomas' NHS Foundation Trust, London, UK
- Chelsea & Westminster Hospital, London, UK
- National Heart & Lung Institute, Imperial College London, London, UK
| | - Emmanouil Karteris
- College of Health, Medicine and Life Sciences, Brunel University London, London, UK
| | - Jan L. Robertus
- Royal Brompton and Harefield Hospitals, Guy's and St Thomas' NHS Foundation Trust, London, UK
- National Heart & Lung Institute, Imperial College London, London, UK
| | | | - Michal Rosen-Zvi
- IBM Research Haifa, Haifa, Israel
- Faculty of Medicine, The Hebrew University of Jerusalem, Jerusalem, Israel
| |
Collapse
|
148
|
Martin-Noguerol T, Luna A. External validation of AI algorithms in breast radiology: the last healthcare security checkpoint? Quant Imaging Med Surg 2021; 11:2888-2892. [PMID: 34079749 DOI: 10.21037/qims-20-1409] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/03/2023]
Affiliation(s)
| | - Antonio Luna
- Radiology Department, HTmédica, Clinica Las Nieves, Jaén, Spain
| |
Collapse
|
149
|
Meyer P, Saez-Rodriguez J. Advances in systems biology modeling: 10 years of crowdsourcing DREAM challenges. Cell Syst 2021; 12:636-653. [PMID: 34139170 DOI: 10.1016/j.cels.2021.05.015] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2020] [Revised: 03/29/2021] [Accepted: 05/18/2021] [Indexed: 02/07/2023]
Abstract
Computational and mathematical models are key to obtain a system-level understanding of biological processes, but their limitations have to be clearly defined to allow their proper application and interpretation. Crowdsourced benchmarks in the form of challenges provide an unbiased assessment of methods, and for the past decade, the Dialogue for Reverse Engineering Assessment and Methods (DREAM) organized more than 15 systems biology challenges. From transcription factor binding to dynamical network models, from signaling networks to gene regulation, from whole-cell models to cell-lineage reconstruction, and from single-cell positioning in a tissue to drug combinations and cell survival, the breadth is broad. To celebrate the 5-year anniversary of Cell Systems, we review the genesis of these systems biology challenges and discuss how interlocking the forward- and reverse-modeling paradigms allows to push the rim of systems biology. This approach will persist for systems levels approaches in biology and medicine.
Collapse
Affiliation(s)
- Pablo Meyer
- IBM T.J. Watson Research Center, Yorktown Heights, NY, USA.
| | - Julio Saez-Rodriguez
- Institute for Computational Biomedicine, Heidelberg University Hospital and Heidelberg University, Faculty of Medicine, Bioquant, Heidelberg 69120, Germany
| |
Collapse
|
150
|
Grimm LJ. Radiomics: A Primer for Breast Radiologists. JOURNAL OF BREAST IMAGING 2021; 3:276-287. [PMID: 38424774 DOI: 10.1093/jbi/wbab014] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2020] [Indexed: 03/02/2024]
Abstract
Radiomics has a long-standing history in breast imaging with computer-aided detection (CAD) for screening mammography developed in the late 20th century. Although conventional CAD had widespread adoption, the clinical benefits for experienced breast radiologists were debatable due to high false-positive marks and subsequent increased recall rates. The dramatic growth in recent years of artificial intelligence-based analysis, including machine learning and deep learning, has provided numerous opportunities for improved modern radiomics work in breast imaging. There has been extensive radiomics work in mammography, digital breast tomosynthesis, MRI, ultrasound, PET-CT, and combined multimodality imaging. Specific radiomics outcomes of interest have been diverse, including CAD, prediction of response to neoadjuvant therapy, lesion classification, and survival, among other outcomes. Additionally, the radiogenomics subfield that correlates radiomics features with genetics has been very proliferative, in parallel with the clinical validation of breast cancer molecular subtypes and gene expression assays. Despite the promise of radiomics, there are important challenges related to image normalization, limited large unbiased data sets, and lack of external validation. Much of the radiomics work to date has been exploratory using single-institution retrospective series for analysis, but several promising lines of investigation have made the leap to clinical practice with commercially available products. As a result, breast radiologists will increasingly be incorporating radiomics-based tools into their daily practice in the near future. Therefore, breast radiologists must have a broad understanding of the scope, applications, and limitations of radiomics work.
Collapse
Affiliation(s)
- Lars J Grimm
- Duke University, Department of Radiology, Durham, NC, USA
| |
Collapse
|