1
|
Yoon SY, Lee KS, Bezuidenhout AF, Kruskal JB. Spectrum of Cognitive Biases in Diagnostic Radiology. Radiographics 2024; 44:e230059. [PMID: 38843094 DOI: 10.1148/rg.230059] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/10/2024]
Abstract
Cognitive biases are systematic thought processes involving the use of a filter of personal experiences and preferences arising from the tendency of the human brain to simplify information processing, especially when taking in vast amounts of data such as from imaging studies. These biases encompass a wide spectrum of thought processes and frequently overlap in their concepts, with multiple biases usually in operation when interpretive and perceptual errors occur in radiology. The authors review the gamut of cognitive biases that occur in radiology. These biases are organized according to their expected stage of occurrence while the radiologist reads and interprets an imaging study. In addition, the authors propose several additional cognitive biases that have not yet, to their knowledge, been defined in the radiologic literature but are applicable to diagnostic radiology. Case examples are used to illustrate potential biases and their impact, with emergency radiology serving as the clinical paradigm, given the associated high imaging volumes, wide diversity of imaging examinations, and rapid pace, which can further increase a radiologist's reliance on biases and heuristics. Potential strategies to recognize and overcome one's personal biases at each stage of image interpretation are also discussed. Awareness of such biases and their unintended effects on imaging interpretations and patient outcomes may help make radiologists cognizant of their own biases that can result in diagnostic errors. Identification of cognitive bias in departmental and systematic quality improvement practices may represent another tool to prevent diagnostic errors in radiology. ©RSNA, 2024 See the invited commentary by Larson in this issue.
Collapse
Affiliation(s)
- Se-Young Yoon
- From the Department of Radiology, Beth Israel Deaconess Medical Center, 330 Brookline Ave, Boston, MA 02215
| | - Karen S Lee
- From the Department of Radiology, Beth Israel Deaconess Medical Center, 330 Brookline Ave, Boston, MA 02215
| | - Abraham F Bezuidenhout
- From the Department of Radiology, Beth Israel Deaconess Medical Center, 330 Brookline Ave, Boston, MA 02215
| | - Jonathan B Kruskal
- From the Department of Radiology, Beth Israel Deaconess Medical Center, 330 Brookline Ave, Boston, MA 02215
| |
Collapse
|
2
|
Carmichael J, Costanza E, Blandford A, Struyven R, Keane PA, Balaskas K. Diagnostic decisions of specialist optometrists exposed to ambiguous deep-learning outputs. Sci Rep 2024; 14:6775. [PMID: 38514657 PMCID: PMC10958016 DOI: 10.1038/s41598-024-55410-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Accepted: 02/23/2024] [Indexed: 03/23/2024] Open
Abstract
Artificial intelligence (AI) has great potential in ophthalmology. We investigated how ambiguous outputs from an AI diagnostic support system (AI-DSS) affected diagnostic responses from optometrists when assessing cases of suspected retinal disease. Thirty optometrists (15 more experienced, 15 less) assessed 30 clinical cases. For ten, participants saw an optical coherence tomography (OCT) scan, basic clinical information and retinal photography ('no AI'). For another ten, they were also given AI-generated OCT-based probabilistic diagnoses ('AI diagnosis'); and for ten, both AI-diagnosis and AI-generated OCT segmentations ('AI diagnosis + segmentation') were provided. Cases were matched across the three types of presentation and were selected to include 40% ambiguous and 20% incorrect AI outputs. Optometrist diagnostic agreement with the predefined reference standard was lowest for 'AI diagnosis + segmentation' (204/300, 68%) compared to 'AI diagnosis' (224/300, 75% p = 0.010), and 'no Al' (242/300, 81%, p = < 0.001). Agreement with AI diagnosis consistent with the reference standard decreased (174/210 vs 199/210, p = 0.003), but participants trusted the AI more (p = 0.029) with segmentations. Practitioner experience did not affect diagnostic responses (p = 0.24). More experienced participants were more confident (p = 0.012) and trusted the AI less (p = 0.038). Our findings also highlight issues around reference standard definition.
Collapse
Affiliation(s)
- Josie Carmichael
- University College London Interaction Centre (UCLIC), UCL, London, UK.
- Institute of Ophthalmology, NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL, London, UK.
| | - Enrico Costanza
- University College London Interaction Centre (UCLIC), UCL, London, UK
| | - Ann Blandford
- University College London Interaction Centre (UCLIC), UCL, London, UK
| | - Robbert Struyven
- Institute of Ophthalmology, NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL, London, UK
| | - Pearse A Keane
- Institute of Ophthalmology, NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL, London, UK
| | - Konstantinos Balaskas
- Institute of Ophthalmology, NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL, London, UK
| |
Collapse
|
3
|
Kim RY. Radiomics and artificial intelligence for risk stratification of pulmonary nodules: Ready for primetime? Cancer Biomark 2024:CBM230360. [PMID: 38427470 PMCID: PMC11300708 DOI: 10.3233/cbm-230360] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/03/2024]
Abstract
Pulmonary nodules are ubiquitously found on computed tomography (CT) imaging either incidentally or via lung cancer screening and require careful diagnostic evaluation and management to both diagnose malignancy when present and avoid unnecessary biopsy of benign lesions. To engage in this complex decision-making, clinicians must first risk stratify pulmonary nodules to determine what the best course of action should be. Recent developments in imaging technology, computer processing power, and artificial intelligence algorithms have yielded radiomics-based computer-aided diagnosis tools that use CT imaging data including features invisible to the naked human eye to predict pulmonary nodule malignancy risk and are designed to be used as a supplement to routine clinical risk assessment. These tools vary widely in their algorithm construction, internal and external validation populations, intended-use populations, and commercial availability. While several clinical validation studies have been published, robust clinical utility and clinical effectiveness data are not yet currently available. However, there is reason for optimism as ongoing and future studies aim to target this knowledge gap, in the hopes of improving the diagnostic process for patients with pulmonary nodules.
Collapse
|
4
|
Li LT, Haley LC, Boyd AK, Bernstam EV. Technical/Algorithm, Stakeholder, and Society (TASS) barriers to the application of artificial intelligence in medicine: A systematic review. J Biomed Inform 2023; 147:104531. [PMID: 37884177 DOI: 10.1016/j.jbi.2023.104531] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2023] [Revised: 09/14/2023] [Accepted: 10/22/2023] [Indexed: 10/28/2023]
Abstract
INTRODUCTION The use of artificial intelligence (AI), particularly machine learning and predictive analytics, has shown great promise in health care. Despite its strong potential, there has been limited use in health care settings. In this systematic review, we aim to determine the main barriers to successful implementation of AI in healthcare and discuss potential ways to overcome these challenges. METHODS We conducted a literature search in PubMed (1/1/2001-1/1/2023). The search was restricted to publications in the English language, and human study subjects. We excluded articles that did not discuss AI, machine learning, predictive analytics, and barriers to the use of these techniques in health care. Using grounded theory methodology, we abstracted concepts to identify major barriers to AI use in medicine. RESULTS We identified a total of 2,382 articles. After reviewing the 306 included papers, we developed 19 major themes, which we categorized into three levels: the Technical/Algorithm, Stakeholder, and Social levels (TASS). These themes included: Lack of Explainability, Need for Validation Protocols, Need for Standards for Interoperability, Need for Reporting Guidelines, Need for Standardization of Performance Metrics, Lack of Plan for Updating Algorithm, Job Loss, Skills Loss, Workflow Challenges, Loss of Patient Autonomy and Consent, Disturbing the Patient-Clinician Relationship, Lack of Trust in AI, Logistical Challenges, Lack of strategic plan, Lack of Cost-effectiveness Analysis and Proof of Efficacy, Privacy, Liability, Bias and Social Justice, and Education. CONCLUSION We identified 19 major barriers to the use of AI in healthcare and categorized them into three levels: the Technical/Algorithm, Stakeholder, and Social levels (TASS). Future studies should expand on barriers in pediatric care and focus on developing clearly defined protocols to overcome these barriers.
Collapse
Affiliation(s)
- Linda T Li
- Department of Surgery, Division of Pediatric Surgery, Icahn School of Medicine at Mount Sinai, 1 Gustave L. Levy Pl, New York, NY 10029, United States; McWilliams School of Biomedical Informatics at UT Health Houston, 7000 Fannin St, Suite 600, Houston, TX 77030, United States.
| | - Lauren C Haley
- McGovern Medical School at the University of Texas Health Science Center at Houston, 6431 Fannin St, Houston, TX 77030, United States.
| | - Alexandra K Boyd
- McGovern Medical School at the University of Texas Health Science Center at Houston, 6431 Fannin St, Houston, TX 77030, United States.
| | - Elmer V Bernstam
- McWilliams School of Biomedical Informatics at UT Health Houston, 7000 Fannin St, Suite 600, Houston, TX 77030, United States; McGovern Medical School at the University of Texas Health Science Center at Houston, 6431 Fannin St, Houston, TX 77030, United States.
| |
Collapse
|
5
|
Lyell D, Magrabi F. Artificial intelligence in medicine: has the time come to hang up the stethoscope? Intern Med J 2023; 53:1533-1539. [PMID: 37683094 PMCID: PMC10946904 DOI: 10.1111/imj.16216] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2023] [Accepted: 08/08/2023] [Indexed: 09/10/2023]
Abstract
The question of whether the time has come to hang up the stethoscope is bound up in the promises of artificial intelligence (AI), promises that have so far proven difficult to deliver, perhaps because of the mismatch between the technical capability of AI and its use in real-world clinical settings. This perspective argues that it is time to move away from discussing the generalised promise of disembodied AI and focus on specifics. We need to focus on how the computational method underlying AI, i.e. machine learning (ML), is embedded into tools, how those tools contribute to clinical tasks and decisions and to what extent they can be relied on. Accordingly, we pose four questions that must be asked to make the discussion real and to understand how ML tools contribute to health care: (1) What does the ML algorithm do? (2) How is output of the ML algorithm used in clinical tools? (3) What does the ML tool contribute to clinical tasks or decisions? (4) Can clinicians act or rely on the ML tool? Two exemplar ML tools are examined to show how these questions can be used to better understand the role of ML in supporting clinical tasks and decisions. Ultimately, ML is just a fancy method of automation. We show that it is useful in automating specific and narrowly defined clinical tasks but likely incapable of automating the full gamut of decisions and tasks performed by clinicians.
Collapse
Affiliation(s)
- David Lyell
- Centre for Health InformaticsAustralian Institute of Health Innovation, Macquarie UniversitySydneyNew South WalesAustralia
| | - Farah Magrabi
- Centre for Health InformaticsAustralian Institute of Health Innovation, Macquarie UniversitySydneyNew South WalesAustralia
| |
Collapse
|
6
|
Eltawil FA, Atalla M, Boulos E, Amirabadi A, Tyrrell PN. Analyzing Barriers and Enablers for the Acceptance of Artificial Intelligence Innovations into Radiology Practice: A Scoping Review. Tomography 2023; 9:1443-1455. [PMID: 37624108 PMCID: PMC10459931 DOI: 10.3390/tomography9040115] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2023] [Revised: 07/23/2023] [Accepted: 07/26/2023] [Indexed: 08/26/2023] Open
Abstract
OBJECTIVES This scoping review was conducted to determine the barriers and enablers associated with the acceptance of artificial intelligence/machine learning (AI/ML)-enabled innovations into radiology practice from a physician's perspective. METHODS A systematic search was performed using Ovid Medline and Embase. Keywords were used to generate refined queries with the inclusion of computer-aided diagnosis, artificial intelligence, and barriers and enablers. Three reviewers assessed the articles, with a fourth reviewer used for disagreements. The risk of bias was mitigated by including both quantitative and qualitative studies. RESULTS An electronic search from January 2000 to 2023 identified 513 studies. Twelve articles were found to fulfill the inclusion criteria: qualitative studies (n = 4), survey studies (n = 7), and randomized controlled trials (RCT) (n = 1). Among the most common barriers to AI implementation into radiology practice were radiologists' lack of acceptance and trust in AI innovations; a lack of awareness, knowledge, and familiarity with the technology; and perceived threat to the professional autonomy of radiologists. The most important identified AI implementation enablers were high expectations of AI's potential added value; the potential to decrease errors in diagnosis; the potential to increase efficiency when reaching a diagnosis; and the potential to improve the quality of patient care. CONCLUSIONS This scoping review found that few studies have been designed specifically to identify barriers and enablers to the acceptance of AI in radiology practice. The majority of studies have assessed the perception of AI replacing radiologists, rather than other barriers or enablers in the adoption of AI. To comprehensively evaluate the potential advantages and disadvantages of integrating AI innovations into radiology practice, gathering more robust research evidence on stakeholder perspectives and attitudes is essential.
Collapse
Affiliation(s)
- Fatma A. Eltawil
- Department of Medical Imaging, University of Toronto, Toronto, ON M5S 1A1, Canada; (F.A.E.); (M.A.); (E.B.)
| | - Michael Atalla
- Department of Medical Imaging, University of Toronto, Toronto, ON M5S 1A1, Canada; (F.A.E.); (M.A.); (E.B.)
| | - Emily Boulos
- Department of Medical Imaging, University of Toronto, Toronto, ON M5S 1A1, Canada; (F.A.E.); (M.A.); (E.B.)
| | - Afsaneh Amirabadi
- Diagnostic Imaging Department, The Hospital for Sick Children, Toronto, ON M5G 1E8, Canada;
| | - Pascal N. Tyrrell
- Department of Medical Imaging, University of Toronto, Toronto, ON M5S 1A1, Canada; (F.A.E.); (M.A.); (E.B.)
- Department of Statistical Sciences, University of Toronto, Toronto, ON M5G 1Z5, Canada
- Institute of Medical Science, University of Toronto, Toronto, ON M5S 1A8, Canada
| |
Collapse
|
7
|
Lee JH, Hong H, Nam G, Hwang EJ, Park CM. Effect of Human-AI Interaction on Detection of Malignant Lung Nodules on Chest Radiographs. Radiology 2023; 307:e222976. [PMID: 37367443 DOI: 10.1148/radiol.222976] [Citation(s) in RCA: 20] [Impact Index Per Article: 20.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/28/2023]
Abstract
Background The factors affecting radiologists' diagnostic determinations in artificial intelligence (AI)-assisted image reading remain underexplored. Purpose To assess how AI diagnostic performance and reader characteristics influence detection of malignant lung nodules during AI-assisted reading of chest radiographs. Materials and Methods This retrospective study consisted of two reading sessions from April 2021 to June 2021. Based on the first session without AI assistance, 30 readers were assigned into two groups with equivalent areas under the free-response receiver operating characteristic curve (AUFROCs). In the second session, each group reinterpreted radiographs assisted by either a high or low accuracy AI model (blinded to the fact that two different AI models were used). Reader performance for detecting lung cancer and reader susceptibility (changing the original reading following the AI suggestion) were compared. A generalized linear mixed model was used to identify the factors influencing AI-assisted detection performance, including readers' attitudes and experiences of AI and Grit score. Results Of the 120 chest radiographs assessed, 60 were obtained in patients with lung cancer (mean age, 67 years ± 12 [SD]; 32 male; 63 cancers) and 60 in controls (mean age, 67 years ± 12; 36 male). Readers included 20 thoracic radiologists (5-18 years of experience) and 10 radiology residents (2-3 years of experience). Use of the high accuracy AI model improved readers' detection performance to a greater extent than use of the low accuracy AI model (area under the receiver operating characteristic curve, 0.77 to 0.82 vs 0.75 to 0.75; AUFROC, 0.71 to 0.79 vs 0.7 to 0.72). Readers who used the high accuracy AI showed a higher susceptibility (67%, 224 of 334 cases) to changing their diagnosis based on the AI suggestions than those using the low accuracy AI (59%, 229 of 386 cases). Accurate readings at the first session, correct AI suggestions, high accuracy Al, and diagnostic difficulty were associated with accurate AI-assisted readings, but readers' characteristics were not. Conclusion An AI model with high diagnostic accuracy led to improved performance of radiologists in detecting lung cancer on chest radiographs and increased radiologists' susceptibility to AI suggestions. © RSNA, 2023 Supplemental material is available for this article.
Collapse
Affiliation(s)
- Jong Hyuk Lee
- From the Department of Radiology (J.H.L., E.J.H., C.M.P.) and Medical Research Collaborating Center (H.H.), Seoul National University Hospital, Seoul, Korea; Lunit, Seoul, Korea (G.N.); Institute of Medical and Biological Engineering and Institute of Radiation Medicine, Seoul National University Medical Research Center, Seoul, Korea (C.M.P.); and Department of Radiology, Seoul National University College of Medicine, 101 Daehak-ro, Jongno-gu, Seoul 03080, Korea (C.M.P.)
| | - Hyunsook Hong
- From the Department of Radiology (J.H.L., E.J.H., C.M.P.) and Medical Research Collaborating Center (H.H.), Seoul National University Hospital, Seoul, Korea; Lunit, Seoul, Korea (G.N.); Institute of Medical and Biological Engineering and Institute of Radiation Medicine, Seoul National University Medical Research Center, Seoul, Korea (C.M.P.); and Department of Radiology, Seoul National University College of Medicine, 101 Daehak-ro, Jongno-gu, Seoul 03080, Korea (C.M.P.)
| | - Gunhee Nam
- From the Department of Radiology (J.H.L., E.J.H., C.M.P.) and Medical Research Collaborating Center (H.H.), Seoul National University Hospital, Seoul, Korea; Lunit, Seoul, Korea (G.N.); Institute of Medical and Biological Engineering and Institute of Radiation Medicine, Seoul National University Medical Research Center, Seoul, Korea (C.M.P.); and Department of Radiology, Seoul National University College of Medicine, 101 Daehak-ro, Jongno-gu, Seoul 03080, Korea (C.M.P.)
| | - Eui Jin Hwang
- From the Department of Radiology (J.H.L., E.J.H., C.M.P.) and Medical Research Collaborating Center (H.H.), Seoul National University Hospital, Seoul, Korea; Lunit, Seoul, Korea (G.N.); Institute of Medical and Biological Engineering and Institute of Radiation Medicine, Seoul National University Medical Research Center, Seoul, Korea (C.M.P.); and Department of Radiology, Seoul National University College of Medicine, 101 Daehak-ro, Jongno-gu, Seoul 03080, Korea (C.M.P.)
| | - Chang Min Park
- From the Department of Radiology (J.H.L., E.J.H., C.M.P.) and Medical Research Collaborating Center (H.H.), Seoul National University Hospital, Seoul, Korea; Lunit, Seoul, Korea (G.N.); Institute of Medical and Biological Engineering and Institute of Radiation Medicine, Seoul National University Medical Research Center, Seoul, Korea (C.M.P.); and Department of Radiology, Seoul National University College of Medicine, 101 Daehak-ro, Jongno-gu, Seoul 03080, Korea (C.M.P.)
| |
Collapse
|
8
|
Wawer Matos PA, Reimer RP, Rokohl AC, Caldeira L, Heindl LM, Große Hokamp N. Artificial Intelligence in Ophthalmology - Status Quo and Future Perspectives. Semin Ophthalmol 2023; 38:226-237. [PMID: 36356300 DOI: 10.1080/08820538.2022.2139625] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
Abstract
Artificial intelligence (AI) is an emerging technology in healthcare and holds the potential to disrupt many arms in medical care. In particular, disciplines using medical imaging modalities, including e.g. radiology but ophthalmology as well, are already confronted with a wide variety of AI implications. In ophthalmologic research, AI has demonstrated promising results limited to specific diseases and imaging tools, respectively. Yet, implementation of AI in clinical routine is not widely spread due to availability, heterogeneity in imaging techniques and AI methods. In order to describe the status quo, this narrational review provides a brief introduction to AI ("what the ophthalmologist needs to know"), followed by an overview of different AI-based applications in ophthalmology and a discussion on future challenges.Abbreviations: Age-related macular degeneration, AMD; Artificial intelligence, AI; Anterior segment OCT, AS-OCT; Coronary artery calcium score, CACS; Convolutional neural network, CNN; Deep convolutional neural network, DCNN; Diabetic retinopathy, DR; Machine learning, ML; Optical coherence tomography, OCT; Retinopathy of prematurity, ROP; Support vector machine, SVM; Thyroid-associated ophthalmopathy, TAO.
Collapse
Affiliation(s)
| | - Robert P Reimer
- Department of Diagnostic and Interventional Radiology, University Hospital of Cologne, Köln, Germany
| | - Alexander C Rokohl
- Department of Ophthalmology, University Hospital of Cologne, Köln, Germany
| | - Liliana Caldeira
- Department of Diagnostic and Interventional Radiology, University Hospital of Cologne, Köln, Germany
| | - Ludwig M Heindl
- Department of Ophthalmology, University Hospital of Cologne, Köln, Germany
| | - Nils Große Hokamp
- Department of Diagnostic and Interventional Radiology, University Hospital of Cologne, Köln, Germany
| |
Collapse
|
9
|
Filler G, Gipson DS, Iyamuremye D, Díaz González de Ferris ME. Artificial Intelligence in Pediatric Nephrology-A Call for Action. ADVANCES IN KIDNEY DISEASE AND HEALTH 2023; 30:17-24. [PMID: 36723276 DOI: 10.1053/j.akdh.2022.11.001] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/18/2022] [Revised: 10/24/2022] [Accepted: 11/07/2022] [Indexed: 12/24/2022]
Abstract
Artificial intelligence is playing an increasingly important role in many fields of clinical care to assist health care providers in patient management. In adult-focused nephrology, artificial intelligence is beginning to be used to improve clinical care, hemodialysis prescriptions, and follow-up of transplant recipients. This article provides an overview of medical artificial intelligence applications relevant to pediatric nephrology. We describe the core concepts of artificial intelligence and machine learning and cover the basics of neural networks and deep learning. We also discuss some examples for clinical applications of artificial intelligence in pediatric nephrology, including neonatal kidney function, early recognition of acute kidney injury, renally cleared drug dosing, intrapatient variability, urinary tract infection workup in infancy, and longitudinal disease progression. Furthermore, we consider the future of artificial intelligence in clinical pediatric nephrology and its potential impact on medical practice and address the ethical issues artificial intelligence raises in terms of clinical decision-making, health care provider-patient relationship, patient privacy, and data collection. This article also represents a call for action involving those of us striving to provide optimal services for children, adolescents, and young adults with chronic conditions.
Collapse
Affiliation(s)
- Guido Filler
- Division of Pediatric Nephrology, Departments of Paediatrics, Western University, London, Ontario, Canada; Departments of Medicine, Western University, London, Ontario, Canada; Department of Pathology and Laboratory Medicine, Western University, London, Ontario, Canada.
| | - Debbie S Gipson
- Department of Pediatrics, University of Michigan, Ann Arbor, Michigan
| | | | | |
Collapse
|
10
|
Monteith S, Glenn T, Geddes J, Whybrow PC, Achtyes E, Bauer M. Expectations for Artificial Intelligence (AI) in Psychiatry. Curr Psychiatry Rep 2022; 24:709-721. [PMID: 36214931 PMCID: PMC9549456 DOI: 10.1007/s11920-022-01378-5] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 09/15/2022] [Indexed: 01/29/2023]
Abstract
PURPOSE OF REVIEW Artificial intelligence (AI) is often presented as a transformative technology for clinical medicine even though the current technology maturity of AI is low. The purpose of this narrative review is to describe the complex reasons for the low technology maturity and set realistic expectations for the safe, routine use of AI in clinical medicine. RECENT FINDINGS For AI to be productive in clinical medicine, many diverse factors that contribute to the low maturity level need to be addressed. These include technical problems such as data quality, dataset shift, black-box opacity, validation and regulatory challenges, and human factors such as a lack of education in AI, workflow changes, automation bias, and deskilling. There will also be new and unanticipated safety risks with the introduction of AI. The solutions to these issues are complex and will take time to discover, develop, validate, and implement. However, addressing the many problems in a methodical manner will expedite the safe and beneficial use of AI to augment medical decision making in psychiatry.
Collapse
Affiliation(s)
- Scott Monteith
- Michigan State University College of Human Medicine, Traverse City Campus, Traverse City, MI, 49684, USA.
| | - Tasha Glenn
- ChronoRecord Association, Fullerton, CA, USA
| | - John Geddes
- Department of Psychiatry, University of Oxford, Warneford Hospital, Oxford, UK
| | - Peter C Whybrow
- Department of Psychiatry and Biobehavioral Sciences, Semel Institute for Neuroscience and Human Behavior, University of California Los Angeles (UCLA), Los Angeles, CA, USA
| | - Eric Achtyes
- Michigan State University College of Human Medicine, Grand Rapids, MI, 49684, USA
- Network180, Grand Rapids, MI, USA
| | - Michael Bauer
- Department of Psychiatry and Psychotherapy, University Hospital Carl Gustav Carus Medical Faculty, Technische Universität Dresden, Dresden, Germany
| |
Collapse
|
11
|
Lahmi L, Mamzer MF, Burgun A, Durdux C, Bibault JE. Ethical Aspects of Artificial Intelligence in Radiation Oncology. Semin Radiat Oncol 2022; 32:442-448. [DOI: 10.1016/j.semradonc.2022.06.013] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
12
|
Marti-Bonmati L, Koh DM, Riklund K, Bobowicz M, Roussakis Y, Vilanova JC, Fütterer JJ, Rimola J, Mallol P, Ribas G, Miguel A, Tsiknakis M, Lekadir K, Tsakou G. Considerations for artificial intelligence clinical impact in oncologic imaging: an AI4HI position paper. Insights Imaging 2022; 13:89. [PMID: 35536446 PMCID: PMC9091068 DOI: 10.1186/s13244-022-01220-9] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2021] [Accepted: 04/07/2022] [Indexed: 01/12/2023] Open
Abstract
To achieve clinical impact in daily oncological practice, emerging AI-based cancer imaging research needs to have clearly defined medical focus, AI methods, and outcomes to be estimated. AI-supported cancer imaging should predict major relevant clinical endpoints, aiming to extract associations and draw inferences in a fair, robust, and trustworthy way. AI-assisted solutions as medical devices, developed using multicenter heterogeneous datasets, should be targeted to have an impact on the clinical care pathway. When designing an AI-based research study in oncologic imaging, ensuring clinical impact in AI solutions requires careful consideration of key aspects, including target population selection, sample size definition, standards, and common data elements utilization, balanced dataset splitting, appropriate validation methodology, adequate ground truth, and careful selection of clinical endpoints. Endpoints may be pathology hallmarks, disease behavior, treatment response, or patient prognosis. Ensuring ethical, safety, and privacy considerations are also mandatory before clinical validation is performed. The Artificial Intelligence for Health Imaging (AI4HI) Clinical Working Group has discussed and present in this paper some indicative Machine Learning (ML) enabled decision-support solutions currently under research in the AI4HI projects, as well as the main considerations and requirements that AI solutions should have from a clinical perspective, which can be adopted into clinical practice. If effectively designed, implemented, and validated, cancer imaging AI-supported tools will have the potential to revolutionize the field of precision medicine in oncology.
Collapse
Affiliation(s)
- Luis Marti-Bonmati
- Radiology Department and Biomedical Imaging Research Group (GIBI230), La Fe Polytechnics and University Hospital and Health Research Institute, Valencia, Spain.
| | - Dow-Mu Koh
- Department of Radiology, Royal Marsden Hospital and Division of Radiotherapy and Imaging, Institute of Cancer Research, London, UK.,Department of Radiology, The Royal Marsden NHS Trust, London, UK
| | - Katrine Riklund
- Department of Radiation Sciences, Diagnostic Radiology, Umeå University, 901 85, Umeå, Sweden
| | - Maciej Bobowicz
- 2nd Department of Radiology, Medical University of Gdansk, 17 Smoluchowskiego Str, 80-214, Gdansk, Poland
| | - Yiannis Roussakis
- Department of Medical Physics, German Oncology Center, 4108, Limassol, Cyprus
| | - Joan C Vilanova
- Department of Radiology, Clínica Girona, Institute of Diagnostic Imaging (IDI)-Girona, Faculty of Medicine, University of Girona, Girona, Spain
| | - Jurgen J Fütterer
- Department of Radiology and Nuclear Medicine, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Jordi Rimola
- CIBERehd, Barcelona Clinic Liver Cancer (BCLC) Group, Department of Radiology, Hospital Clínic, University of Barcelona, Barcelona, Spain
| | - Pedro Mallol
- Radiology Department and Biomedical Imaging Research Group (GIBI230), La Fe Polytechnics and University Hospital and Health Research Institute, Valencia, Spain
| | - Gloria Ribas
- Radiology Department and Biomedical Imaging Research Group (GIBI230), La Fe Polytechnics and University Hospital and Health Research Institute, Valencia, Spain
| | - Ana Miguel
- Radiology Department and Biomedical Imaging Research Group (GIBI230), La Fe Polytechnics and University Hospital and Health Research Institute, Valencia, Spain
| | - Manolis Tsiknakis
- Foundation for Research and Technology Hellas, Institute of Computer Science, Computational Biomedicine Lab (CBML), FORTH-ICS Heraklion, Crete, Greece
| | - Karim Lekadir
- Departament de Matemàtiques and Informàtica, Artificial Intelligence in Medicine Lab (BCN-AIM), Universitat de Barcelona, Barcelona, Spain
| | - Gianna Tsakou
- Maggioli S.P.A., Research and Development Lab, Athens, Greece
| |
Collapse
|
13
|
Ploug T, Holm S. Right to Contest AI Diagnostics. Artif Intell Med 2022. [DOI: 10.1007/978-3-030-64573-1_267] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
14
|
Kämmer JE, Schauber SK, Hautz SC, Stroben F, Hautz WE. Differential diagnosis checklists reduce diagnostic error differentially: A randomised experiment. MEDICAL EDUCATION 2021; 55:1172-1182. [PMID: 34291481 PMCID: PMC9290564 DOI: 10.1111/medu.14596] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/06/2021] [Accepted: 07/13/2021] [Indexed: 05/30/2023]
Abstract
INTRODUCTION Wrong and missed diagnoses contribute substantially to medical error. Can a prompt to generate alternative diagnoses (prompt) or a differential diagnosis checklist (DDXC) increase diagnostic accuracy? How do these interventions affect the diagnostic process and self-monitoring? METHODS Advanced medical students (N = 90) were randomly assigned to one of four conditions to complete six computer-based patient cases: group 1 (prompt) was instructed to write down all diagnoses they considered while acquiring diagnostic test results and to finally rank them. Groups 2 and 3 received the same instruction plus a list of 17 differential diagnoses for the chief complaint of the patient. For half of the cases, the DDXC contained the correct diagnosis (DDXC+), and for the other half, it did not (DDXC-; counterbalanced). Group 4 (control) was only instructed to indicate their final diagnosis. Mixed-effects models were used to analyse results. RESULTS Students using a DDXC that contained the correct diagnosis had better diagnostic accuracy, mean (standard deviation), 0.75 (0.44), compared to controls without a checklist, 0.49 (0.50), P < 0.001, but those using a DDXC that did not contain the correct diagnosis did slightly worse, 0.43 (0.50), P = 0.602. The number and relevance of diagnostic tests acquired were not affected by condition, nor was self-monitoring. However, participants spent more time on a case in the DDXC-, 4:20 min (2:36), P ≤ 0.001, and DDXC+ condition, 3:52 min (2:09), than in the control condition, 2:59 min (1:44), P ≤ 0.001. DISCUSSION Being provided a list of possible diagnoses improves diagnostic accuracy compared with a prompt to create a differential diagnosis list, if the provided list contains the correct diagnosis. However, being provided a diagnosis list without the correct diagnosis did not improve and might have slightly reduced diagnostic accuracy. Interventions neither affected information gathering nor self-monitoring.
Collapse
Affiliation(s)
- Juliane E. Kämmer
- Department of Emergency Medicine, Inselspital University HospitalUniversity of BernBernSwitzerland
- Center for Adaptive Rationality (ARC)Max Planck Institute for Human DevelopmentBerlinGermany
| | - Stefan K. Schauber
- Centre for Health Sciences Education, Faculty of MedicineUniversity of OsloOsloNorway
| | - Stefanie C. Hautz
- Department of Emergency Medicine, Inselspital University HospitalUniversity of BernBernSwitzerland
| | - Fabian Stroben
- Department of Anesthesiology and Operative Intensive Care Medicine (CBF), Charité – Universitätsmedizin BerlinHumboldt University of BerlinBerlinGermany
| | - Wolf E. Hautz
- Department of Emergency Medicine, Inselspital University HospitalUniversity of BernBernSwitzerland
| |
Collapse
|
15
|
Abdullah YI, Schuman JS, Shabsigh R, Caplan A, Al-Aswad LA. Ethics of Artificial Intelligence in Medicine and Ophthalmology. Asia Pac J Ophthalmol (Phila) 2021; 10:289-298. [PMID: 34383720 PMCID: PMC9167644 DOI: 10.1097/apo.0000000000000397] [Citation(s) in RCA: 24] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022] Open
Abstract
BACKGROUND This review explores the bioethical implementation of artificial intelligence (AI) in medicine and in ophthalmology. AI, which was first introduced in the 1950s, is defined as "the machine simulation of human mental reasoning, decision making, and behavior". The increased power of computing, expansion of storage capacity, and compilation of medical big data helped the AI implementation surge in medical practice and research. Ophthalmology is a leading medical specialty in applying AI in screening, diagnosis, and treatment. The first Food and Drug Administration approved autonomous diagnostic system served to diagnose and classify diabetic retinopathy. Other ophthalmic conditions such as age-related macular degeneration, glaucoma, retinopathy of prematurity, and congenital cataract, among others, implemented AI too. PURPOSE To review the contemporary literature of the bioethical issues of AI in medicine and ophthalmology, classify ethical issues in medical AI, and suggest possible standardizations of ethical frameworks for AI implementation. METHODS Keywords were searched on Google Scholar and PubMed between October 2019 and April 2020. The results were reviewed, cross-referenced, and summarized. A total of 284 references including articles, books, book chapters, and regulatory reports and statements were reviewed, and those that were relevant were cited in the paper. RESULTS Most sources that studied the use of AI in medicine explored the ethical aspects. Bioethical challenges of AI implementation in medicine were categorized into 6 main categories. These include machine training ethics, machine accuracy ethics, patient-related ethics, physician-related ethics, shared ethics, and roles of regulators. CONCLUSIONS There are multiple stakeholders in the ethical issues surrounding AI in medicine and ophthalmology. Attention to the various aspects of ethics related to AI is important especially with the expanding use of AI. Solutions of ethical problems are envisioned to be multifactorial.
Collapse
Affiliation(s)
| | - Joel S Schuman
- Department of Ophthalmology, NYU Langone Health, NYU Grossman School of Medicine, New York, NY
- Department of Biomedical Engineering, NYU Tandon School of Engineering, Brooklyn, NY
- Department of Electrical and Computer Engineering, NYU Tandon School of Engineering, Brooklyn, NY
- Department of Physiology and Neuroscience, NYU Langone Health, NYU Grossman School of Medicine, New York, NY
- Center for Neural Science, NYU College of Arts and Science, New York, NY
| | - Ridwan Shabsigh
- SBH Health System and Weill Cornell Medical College, New York, NY
| | - Arthur Caplan
- Department of Population Health, NYU Langone Health, NYU Grossman School of Medicine, New York, NY
| | - Lama A Al-Aswad
- Department of Ophthalmology, NYU Langone Health, NYU Grossman School of Medicine, New York, NY
- Department of Population Health, NYU Langone Health, NYU Grossman School of Medicine, New York, NY
| |
Collapse
|
16
|
Gaube S, Suresh H, Raue M, Merritt A, Berkowitz SJ, Lermer E, Coughlin JF, Guttag JV, Colak E, Ghassemi M. Do as AI say: susceptibility in deployment of clinical decision-aids. NPJ Digit Med 2021; 4:31. [PMID: 33608629 PMCID: PMC7896064 DOI: 10.1038/s41746-021-00385-9] [Citation(s) in RCA: 122] [Impact Index Per Article: 40.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/21/2020] [Accepted: 01/07/2021] [Indexed: 02/07/2023] Open
Abstract
Artificial intelligence (AI) models for decision support have been developed for clinical settings such as radiology, but little work evaluates the potential impact of such systems. In this study, physicians received chest X-rays and diagnostic advice, some of which was inaccurate, and were asked to evaluate advice quality and make diagnoses. All advice was generated by human experts, but some was labeled as coming from an AI system. As a group, radiologists rated advice as lower quality when it appeared to come from an AI system; physicians with less task-expertise did not. Diagnostic accuracy was significantly worse when participants received inaccurate advice, regardless of the purported source. This work raises important considerations for how advice, AI and non-AI, should be deployed in clinical environments.
Collapse
Affiliation(s)
- Susanne Gaube
- Department of Psychology, University of Regensburg, Regensburg, Germany. .,MIT AgeLab, Massachusetts Institute of Technology, Cambridge, MA, USA.
| | - Harini Suresh
- MIT Computer Science & Artificial Intelligence Lab, Massachusetts Institute of Technology, Cambridge, MA, USA.
| | - Martina Raue
- MIT AgeLab, Massachusetts Institute of Technology, Cambridge, MA, USA
| | | | - Seth J Berkowitz
- Department of Radiology, Beth Israel Deaconess Medical Center, Boston, MA, USA
| | - Eva Lermer
- LMU Center for Leadership and People Management, LMU Munich, Munich, Germany.,FOM University of Applied Sciences for Economics & Management, Munich, Germany
| | - Joseph F Coughlin
- MIT AgeLab, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - John V Guttag
- MIT Computer Science & Artificial Intelligence Lab, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Errol Colak
- Li Ka Shing Knowledge Institute, St. Michael's Hospital, Toronto, Canada.,Department of Medical Imaging, University of Toronto, Toronto, Canada
| | - Marzyeh Ghassemi
- Departments of Computer Science and Medicine, University of Toronto, Toronto, Canada.,Vector Institute, Toronto, Canada
| |
Collapse
|
17
|
Pot M, Kieusseyan N, Prainsack B. Not all biases are bad: equitable and inequitable biases in machine learning and radiology. Insights Imaging 2021; 12:13. [PMID: 33564955 PMCID: PMC7872878 DOI: 10.1186/s13244-020-00955-7] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/10/2020] [Accepted: 12/14/2020] [Indexed: 11/10/2022] Open
Abstract
The application of machine learning (ML) technologies in medicine generally but also in radiology more specifically is hoped to improve clinical processes and the provision of healthcare. A central motivation in this regard is to advance patient treatment by reducing human error and increasing the accuracy of prognosis, diagnosis and therapy decisions. There is, however, also increasing awareness about bias in ML technologies and its potentially harmful consequences. Biases refer to systematic distortions of datasets, algorithms, or human decision making. These systematic distortions are understood to have negative effects on the quality of an outcome in terms of accuracy, fairness, or transparency. But biases are not only a technical problem that requires a technical solution. Because they often also have a social dimension, the 'distorted' outcomes they yield often have implications for equity. This paper assesses different types of biases that can emerge within applications of ML in radiology, and discusses in what cases such biases are problematic. Drawing upon theories of equity in healthcare, we argue that while some biases are harmful and should be acted upon, others might be unproblematic and even desirable-exactly because they can contribute to overcome inequities.
Collapse
Affiliation(s)
- Mirjam Pot
- Department of Political Science, University of Vienna, Austria, Universitätsstraße 7, 1100, Wien, Austria
| | | | - Barbara Prainsack
- Department of Political Science, University of Vienna, Austria, Universitätsstraße 7, 1100, Wien, Austria. .,Department of Global Health and Social Medicine, King's College London, London, UK.
| |
Collapse
|
18
|
Right to Contest AI Diagnostics. Artif Intell Med 2021. [DOI: 10.1007/978-3-030-58080-3_267-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
19
|
What does it mean to provide decision support to a responsible and competent expert? EURO JOURNAL ON DECISION PROCESSES 2020. [DOI: 10.1007/s40070-020-00116-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
20
|
Chong NK, Chu Shan Elaine C, de Korne DF. Creating a Learning Televillage and Automated Digital Child Health Ecosystem. Pediatr Clin North Am 2020; 67:707-724. [PMID: 32650868 DOI: 10.1016/j.pcl.2020.04.016] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
Abstract
This article explores the impact of digital technologies, including telehealth, teleconsultations, wireless devices, and chatbots, in pediatrics. Automated digital health with the Internet of things will allow better collection of real-world data for generation of real-world evidence to improve child health. Artificial intelligence with predictive analytics in turn will drive evidence-based decision-support systems and deliver personalized care to children. This technology creates building blocks for a learning child health and health care ecosystem.
Collapse
Affiliation(s)
- Ng Kee Chong
- Medical Innovation & Care Transformation, Division of Medicine, KK Women's & Children's Hospital, 100 Bukit Timah Road, Singapore 229899, Singapore.
| | - Chew Chu Shan Elaine
- Adolescent Medicine Service, Department of Paediatrics, KK Women's & Children's Hospital, Singapore
| | - Dirk F de Korne
- Medical Innovation & Care Transformation, KK Women's & Children's Hospital, Singapore; Erasmus School of Health Policy & Management, Erasmus University Rotterdam, Netherlands; SVRZ Cares in Zeeland, Middelburg, Netherlands
| |
Collapse
|
21
|
The four dimensions of contestable AI diagnostics - A patient-centric approach to explainable AI. Artif Intell Med 2020; 107:101901. [DOI: 10.1016/j.artmed.2020.101901] [Citation(s) in RCA: 36] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2020] [Revised: 04/30/2020] [Accepted: 06/03/2020] [Indexed: 01/17/2023]
|
22
|
Vollmer S, Mateen BA, Bohner G, Király FJ, Ghani R, Jonsson P, Cumbers S, Jonas A, McAllister KSL, Myles P, Granger D, Birse M, Branson R, Moons KGM, Collins GS, Ioannidis JPA, Holmes C, Hemingway H. Machine learning and artificial intelligence research for patient benefit: 20 critical questions on transparency, replicability, ethics, and effectiveness. BMJ 2020; 368:l6927. [PMID: 32198138 PMCID: PMC11515850 DOI: 10.1136/bmj.l6927] [Citation(s) in RCA: 155] [Impact Index Per Article: 38.8] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 10/22/2019] [Indexed: 12/11/2022]
Affiliation(s)
- Sebastian Vollmer
- Alan Turing Institute, Kings Cross, London, UK
- Departments of Mathematics and Statistics, University of Warwick, Coventry, UK
| | - Bilal A Mateen
- Alan Turing Institute, Kings Cross, London, UK
- Warwick Medical School, University of Warwick, Coventry, UK
- Kings College Hospital, Denmark Hill, London, UK
| | - Gergo Bohner
- Alan Turing Institute, Kings Cross, London, UK
- Departments of Mathematics and Statistics, University of Warwick, Coventry, UK
| | - Franz J Király
- Alan Turing Institute, Kings Cross, London, UK
- Department of Statistical Science, University College London, London, UK
| | | | - Pall Jonsson
- Science Policy and Research, National Institute for Health and Care Excellence, Manchester, UK
| | - Sarah Cumbers
- Health and Social Care Directorate, National Institute for Health and Care Excellence, London, UK
| | - Adrian Jonas
- Data and Analytics Group, National Institute for Health and Care Excellence, London, UK
| | | | - Puja Myles
- Clinical Practice Research Datalink, Medicines and Healthcare products Regulatory Agency, London, UK
| | - David Granger
- Medicines and Healthcare products Regulatory Agency, London, UK
| | - Mark Birse
- Medicines and Healthcare products Regulatory Agency, London, UK
| | - Richard Branson
- Medicines and Healthcare products Regulatory Agency, London, UK
| | - Karel G M Moons
- Julius Centre for Health Sciences and Primary Care, UMC Utrecht, Utrecht University, Utrecht, Netherlands
| | - Gary S Collins
- UK EQUATOR Centre, Centre for Statistics in Medicine, NDORMS, University of Oxford, Oxford, UK
| | - John P A Ioannidis
- Meta-Research Innovation Centre at Stanford, Stanford University, Stanford, CA, USA
| | - Chris Holmes
- Alan Turing Institute, Kings Cross, London, UK
- Department of Statistics, University of Oxford, Oxford OX1 3LB, UK
| | - Harry Hemingway
- Health Data Research UK London, University College London, London, UK
- Institute of Health Informatics, University College London, London, UK
- National Institute for Health Research, University College London Hospitals Biomedical Research Centre, University College London, London, UK
| |
Collapse
|
23
|
Automated prediction of mastitis infection patterns in dairy herds using machine learning. Sci Rep 2020; 10:4289. [PMID: 32152401 PMCID: PMC7062853 DOI: 10.1038/s41598-020-61126-8] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2019] [Accepted: 02/18/2020] [Indexed: 11/29/2022] Open
Abstract
Mastitis in dairy cattle is extremely costly both in economic and welfare terms and is one of the most significant drivers of antimicrobial usage in dairy cattle. A critical step in the prevention of mastitis is the diagnosis of the predominant route of transmission of pathogens into either contagious (CONT) or environmental (ENV), with environmental being further subdivided as transmission during either the nonlactating “dry” period (EDP) or lactating period (EL). Using data from 1000 farms, random forest algorithms were able to replicate the complex herd level diagnoses made by specialist veterinary clinicians with a high degree of accuracy. An accuracy of 98%, positive predictive value (PPV) of 86% and negative predictive value (NPV) of 99% was achieved for the diagnosis of CONT vs ENV (with CONT as a “positive” diagnosis), and an accuracy of 78%, PPV of 76% and NPV of 81% for the diagnosis of EDP vs EL (with EDP as a “positive” diagnosis). An accurate, automated mastitis diagnosis tool has great potential to aid non-specialist veterinary clinicians to make a rapid herd level diagnosis and promptly implement appropriate control measures for an extremely damaging disease in terms of animal health, productivity, welfare and antimicrobial use.
Collapse
|
24
|
Armstrong GW, Lorch AC. A(eye): A Review of Current Applications of Artificial Intelligence and Machine Learning in Ophthalmology. Int Ophthalmol Clin 2020; 60:57-71. [PMID: 31855896 DOI: 10.1097/iio.0000000000000298] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
|
25
|
Lyell D, Magrabi F, Coiera E. Reduced Verification of Medication Alerts Increases Prescribing Errors. Appl Clin Inform 2019; 10:66-76. [PMID: 30699458 DOI: 10.1055/s-0038-1677009] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022] Open
Abstract
OBJECTIVE Clinicians using clinical decision support (CDS) to prescribe medications have an obligation to ensure that prescriptions are safe. One option is to verify the safety of prescriptions if there is uncertainty, for example, by using drug references. Supervisory control experiments in aviation and process control have associated errors, with reduced verification arising from overreliance on decision support. However, it is unknown whether this relationship extends to clinical decision-making. Therefore, we examine whether there is a relationship between verification behaviors and prescribing errors, with and without CDS medication alerts, and whether task complexity mediates this. METHODS A total of 120 students in the final 2 years of a medical degree prescribed medicines for patient scenarios using a simulated electronic prescribing system. CDS (correct, incorrect, and no CDS) and task complexity (low and high) were varied. Outcomes were omission (missed prescribing errors) and commission errors (accepted false-positive alerts). Verification measures were access of drug references and view time percentage of task time. RESULTS Failure to access references for medicines with prescribing errors increased omission errors with no CDS (high-complexity: χ 2(1) = 12.716; p < 0.001) and incorrect CDS (Fisher's exact; low-complexity: p = 0.002; high-complexity: p = 0.001). Failure to access references for false-positive alerts increased commission errors (low-complexity: χ 2(1) = 16.673, p < 0.001; high-complexity: χ 2(1) = 18.690, p < 0.001). Fewer participants accessed relevant references with incorrect CDS compared with no CDS (McNemar; low-complexity: p < 0.001; high-complexity: p < 0.001). Lower view time percentages increased omission (F(3, 361.914) = 4.498; p = 0.035) and commission errors (F(1, 346.223) = 2.712; p = 0.045). View time percentages were lower in CDS-assisted conditions compared with unassisted conditions (F(2, 335.743) = 10.443; p < 0.001). DISCUSSION The presence of CDS reduced verification of prescription safety. When CDS was incorrect, reduced verification was associated with increased prescribing errors. CONCLUSION CDS can be incorrect, and verification provides one mechanism to detect errors. System designers need to facilitate verification without increasing workload or eliminating the benefits of correct CDS.
Collapse
Affiliation(s)
- David Lyell
- Centre for Health Informatics, Australian Institute of Health Innovation, Macquarie University, Sydney, New South Wales, Australia
| | - Farah Magrabi
- Centre for Health Informatics, Australian Institute of Health Innovation, Macquarie University, Sydney, New South Wales, Australia
| | - Enrico Coiera
- Centre for Health Informatics, Australian Institute of Health Innovation, Macquarie University, Sydney, New South Wales, Australia
| |
Collapse
|
26
|
Lyell D, Magrabi F, Coiera E. The Effect of Cognitive Load and Task Complexity on Automation Bias in Electronic Prescribing. HUMAN FACTORS 2018; 60:1008-1021. [PMID: 29939764 DOI: 10.1177/0018720818781224] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
OBJECTIVE Determine the relationship between cognitive load (CL) and automation bias (AB). BACKGROUND Clinical decision support (CDS) for electronic prescribing can improve safety but introduces the risk of AB, where reliance on CDS replaces vigilance in information seeking and processing. We hypothesized high CL generated by high task complexity would increase AB errors. METHOD One hundred twenty medical students prescribed medicines for clinical scenarios using a simulated e-prescribing system in a randomized controlled experiment. Quality of CDS (correct, incorrect, and no CDS) and task complexity (low and high) were varied. CL, omission errors (failure to detect prescribing errors), and commission errors (acceptance of false positive alerts) were measured. RESULTS Increasing complexity from low to high significantly increased CL, F(1, 118) = 71.6, p < .001. CDS reduced CL in high-complexity conditions compared to no CDS, F(2, 117) = 4.72, p = .015. Participants who made omission errors in incorrect and no CDS conditions exhibited lower CL compared to those who did not, F(1, 636.49) = 3.79, p = .023. CONCLUSION Results challenge the notion that AB is triggered by increasing task complexity and associated increases in CL. Omission errors were associated with lower CL, suggesting errors may stem from an insufficient allocation of cognitive resources. APPLICATION This is the first research to examine the relationship between CL and AB. Findings suggest designers and users of CDS systems need to be aware of the risks of AB. Interventions that increase user vigilance and engagement may be beneficial and deserve further investigation.
Collapse
Affiliation(s)
- David Lyell
- Macquarie University, Sydney, New South Wales, Australia
| | - Farah Magrabi
- Macquarie University, Sydney, New South Wales, Australia
| | - Enrico Coiera
- Macquarie University, Sydney, New South Wales, Australia
| |
Collapse
|
27
|
Abstract
Machine learning (ML) has the potential to significantly aid medical practice. However, a recent article highlighted some negative consequences that may arise from using ML decision support in medicine. We argue here that whilst the concerns raised by the authors may be appropriate, they are not specific to ML, and thus the article may lead to an adverse perception about this technique in particular. Whilst ML is not without its limitations like any methodology, a balanced view is needed in order to not hamper its use in potentially enabling better patient care.
Collapse
Affiliation(s)
- Laura McDonald
- Centre for Observational Research and Data Sciences, Bristol-Myers Squibb, Uxbridge, UB8 1DH, UK
| | - Sreeram V Ramagopalan
- Centre for Observational Research and Data Sciences, Bristol-Myers Squibb, Uxbridge, UB8 1DH, UK
| | | | | |
Collapse
|
28
|
Ivlev I, Hickman EN, McDonagh MS, Eden KB. Use of patient decision aids increased younger women's reluctance to begin screening mammography: a systematic review and meta-analysis. J Gen Intern Med 2017; 32:803-812. [PMID: 28289963 PMCID: PMC5481237 DOI: 10.1007/s11606-017-4027-9] [Citation(s) in RCA: 17] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/27/2016] [Revised: 02/07/2017] [Accepted: 02/21/2017] [Indexed: 12/15/2022]
Abstract
BACKGROUND As breast cancer screening guidelines have changed recently, additional investigation is needed to understand changes in women's behavior after using breast cancer screening patient decision aids (BCS-PtDAs) and the potential effect on mammography utilization. This systematic review and meta-analysis sought to evaluate the effect of BCS-PtDAs on changes in women's intentions to undergo screening mammography and whether women deciding to begin or discontinue screening mammography displayed similar changes in screening intentions after using a BCS-PtDA. METHODS We searched Medline, Scopus, PsycINFO, CENTRAL, Health and Psychosocial Instruments, Health Technology Assessment Database, PsycARTICLES, and cited references in eligible papers for randomized controlled trials (RCTs) and observational studies, published through August 24, 2016. The proportions of women who did and not intend to undergo screening and who were uncertain about undergoing screening mammography were pooled, using risk ratios (RR) and random effects. According to the protocol, RCTs or observational studies and any language were considered eligible for systematic review if they included data about women for which shared decision making is recommended. RESULTS We ultimately included six studies with screening intention data for 2040 women. Compared to usual care, the use of BCS-PtDAs in three RCTs resulted in significantly more women deciding not to undergo screening mammography (RR 1.48 [95% CI 1.04-2.13]; P = 0.03), particularly for younger (38-50 years) women (1.77 [1.34-2.34]; P < 0.001). The use of BCS-PtDAs had a non-significant effect on the intentions of older women (69-89 years) to discontinue screening. CONCLUSIONS The use of BCS-PtDAs increased younger women's reluctance to undergo screening for breast cancer. The implementation of such BCS-PtDAs in clinical practice would be expected to result in a 77% increase in the number of younger women (aged 38-50) who do not intend to be screened, and as a consequence, may reduce utilization of screening mammography. REGISTRATION The protocol of this review is registered in the PROSPERO database, #CRD42016036695.
Collapse
Affiliation(s)
- Ilya Ivlev
- Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR, USA.
| | - Erin N Hickman
- Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR, USA
| | - Marian S McDonagh
- Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR, USA
- Pacific Northwest Evidence-based Practice Center, Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR, USA
| | - Karen B Eden
- Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR, USA
- Pacific Northwest Evidence-based Practice Center, Department of Medical Informatics & Clinical Epidemiology, Oregon Health & Science University, Portland, OR, USA
| |
Collapse
|
29
|
Lyell D, Magrabi F, Raban MZ, Pont LG, Baysari MT, Day RO, Coiera E. Automation bias in electronic prescribing. BMC Med Inform Decis Mak 2017; 17:28. [PMID: 28302112 PMCID: PMC5356416 DOI: 10.1186/s12911-017-0425-5] [Citation(s) in RCA: 36] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2016] [Accepted: 03/09/2017] [Indexed: 11/22/2022] Open
Abstract
Background Clinical decision support (CDS) in e-prescribing can improve safety by alerting potential errors, but introduces new sources of risk. Automation bias (AB) occurs when users over-rely on CDS, reducing vigilance in information seeking and processing. Evidence of AB has been found in other clinical tasks, but has not yet been tested with e-prescribing. This study tests for the presence of AB in e-prescribing and the impact of task complexity and interruptions on AB. Methods One hundred and twenty students in the final two years of a medical degree prescribed medicines for nine clinical scenarios using a simulated e-prescribing system. Quality of CDS (correct, incorrect and no CDS) and task complexity (low, low + interruption and high) were varied between conditions. Omission errors (failure to detect prescribing errors) and commission errors (acceptance of false positive alerts) were measured. Results Compared to scenarios with no CDS, correct CDS reduced omission errors by 38.3% (p < .0001, n = 120), 46.6% (p < .0001, n = 70), and 39.2% (p < .0001, n = 120) for low, low + interrupt and high complexity scenarios respectively. Incorrect CDS increased omission errors by 33.3% (p < .0001, n = 120), 24.5% (p < .009, n = 82), and 26.7% (p < .0001, n = 120). Participants made commission errors, 65.8% (p < .0001, n = 120), 53.5% (p < .0001, n = 82), and 51.7% (p < .0001, n = 120). Task complexity and interruptions had no impact on AB. Conclusions This study found evidence of AB omission and commission errors in e-prescribing. Verification of CDS alerts is key to avoiding AB errors. However, interventions focused on this have had limited success to date. Clinicians should remain vigilant to the risks of CDS failures and verify CDS. Electronic supplementary material The online version of this article (doi:10.1186/s12911-017-0425-5) contains supplementary material, which is available to authorized users.
Collapse
Affiliation(s)
- David Lyell
- Centre for Health Informatics, Australian Institute of Health Innovation, Macquarie University, Sydney, NSW, 2109, Australia.
| | - Farah Magrabi
- Centre for Health Informatics, Australian Institute of Health Innovation, Macquarie University, Sydney, NSW, 2109, Australia
| | - Magdalena Z Raban
- Centre for Health Systems and Safety Research, Australian Institute of Health Innovation, Macquarie University, Sydney, NSW, 2109, Australia
| | - L G Pont
- Centre for Health Systems and Safety Research, Australian Institute of Health Innovation, Macquarie University, Sydney, NSW, 2109, Australia
| | - Melissa T Baysari
- Centre for Health Systems and Safety Research, Australian Institute of Health Innovation, Macquarie University, Sydney, NSW, 2109, Australia.,St Vincent's Clinical School, Faculty of Medicine, University of New South Wales, Sydney, Australia
| | - Richard O Day
- St Vincent's Hospital Clinical School and Pharmacology, School of Medical Sciences, Faculty of Medicine, University of New South Wales, Sydney, Australia
| | - Enrico Coiera
- Centre for Health Informatics, Australian Institute of Health Innovation, Macquarie University, Sydney, NSW, 2109, Australia
| |
Collapse
|
30
|
Lyell D, Coiera E. Automation bias and verification complexity: a systematic review. J Am Med Inform Assoc 2017; 24:423-431. [PMID: 27516495 PMCID: PMC7651899 DOI: 10.1093/jamia/ocw105] [Citation(s) in RCA: 93] [Impact Index Per Article: 13.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2016] [Revised: 05/24/2016] [Accepted: 05/27/2016] [Indexed: 01/01/2023] Open
Abstract
INTRODUCTION While potentially reducing decision errors, decision support systems can introduce new types of errors. Automation bias (AB) happens when users become overreliant on decision support, which reduces vigilance in information seeking and processing. Most research originates from the human factors literature, where the prevailing view is that AB occurs only in multitasking environments. OBJECTIVES This review seeks to compare the human factors and health care literature, focusing on the apparent association of AB with multitasking and task complexity. DATA SOURCES EMBASE, Medline, Compendex, Inspec, IEEE Xplore, Scopus, Web of Science, PsycINFO, and Business Source Premiere from 1983 to 2015. STUDY SELECTION Evaluation studies where task execution was assisted by automation and resulted in errors were included. Participants needed to be able to verify automation correctness and perform the task manually. METHODS Tasks were identified and grouped. Task and automation type and presence of multitasking were noted. Each task was rated for its verification complexity. RESULTS Of 890 papers identified, 40 met the inclusion criteria; 6 were in health care. Contrary to the prevailing human factors view, AB was found in single tasks, typically involving diagnosis rather than monitoring, and with high verification complexity. LIMITATIONS The literature is fragmented, with large discrepancies in how AB is reported. Few studies reported the statistical significance of AB compared to a control condition. CONCLUSION AB appears to be associated with the degree of cognitive load experienced in decision tasks, and appears to not be uniquely associated with multitasking. Strategies to minimize AB might focus on cognitive load reduction.
Collapse
Affiliation(s)
- David Lyell
- Centre for Health Informatics, Australian Institute of Health Innovation, Macquarie University, Sydney, Australia
| | - Enrico Coiera
- Centre for Health Informatics, Australian Institute of Health Innovation, Macquarie University, Sydney, Australia
| |
Collapse
|
31
|
Jorritsma W, Cnossen F, van Ooijen PMA. Improving the radiologist-CAD interaction: designing for appropriate trust. Clin Radiol 2014; 70:115-22. [PMID: 25459198 DOI: 10.1016/j.crad.2014.09.017] [Citation(s) in RCA: 30] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2014] [Revised: 09/17/2014] [Accepted: 09/19/2014] [Indexed: 12/25/2022]
Abstract
Computer-aided diagnosis (CAD) has great potential to improve radiologists' diagnostic performance. However, the reported performance of the radiologist-CAD team is lower than what might be expected based on the performance of the radiologist and the CAD system in isolation. This indicates that the interaction between radiologists and the CAD system is not optimal. An important factor in the interaction between humans and automated aids (such as CAD) is trust. Suboptimal performance of the human-automation team is often caused by an inappropriate level of trust in the automation. In this review, we examine the role of trust in the radiologist-CAD interaction and suggest ways to improve the output of the CAD system so that it allows radiologists to calibrate their trust in the CAD system more effectively. Observer studies of the CAD systems show that radiologists often have an inappropriate level of trust in the CAD system. They sometimes under-trust CAD, thereby reducing its potential benefits, and sometimes over-trust it, leading to diagnostic errors they would not have made without CAD. Based on the literature on trust in human-automation interaction and the results of CAD observer studies, we have identified four ways to improve the output of CAD so that it allows radiologists to form a more appropriate level of trust in CAD. Designing CAD systems for appropriate trust is important and can improve the performance of the radiologist-CAD team. Future CAD research and development should acknowledge the importance of the radiologist-CAD interaction, and specifically the role of trust therein, in order to create the perfect artificial partner for the radiologist. This review focuses on the role of trust in the radiologist-CAD interaction. The aim of the review is to encourage CAD developers to design for appropriate trust and thereby improve the performance of the radiologist-CAD team.
Collapse
Affiliation(s)
- W Jorritsma
- Department of Radiology, University of Groningen, University Medical Center Groningen, Hanzeplein 1, 9713 GZ, Groningen, The Netherlands.
| | - F Cnossen
- Institute of Artificial Intelligence and Cognitive Engineering, University of Groningen, Nijenborgh 9, 9747 AG, Groningen, The Netherlands
| | - P M A van Ooijen
- Department of Radiology, University of Groningen, University Medical Center Groningen, Hanzeplein 1, 9713 GZ, Groningen, The Netherlands; Center for Medical Imaging North East Netherlands, Hanzeplein 1, 9713 GZ, Groningen, The Netherlands
| |
Collapse
|
32
|
Fenton JJ, Xing G, Elmore JG, Bang H, Chen SL, Lindfors KK, Baldwin LM. Short-term outcomes of screening mammography using computer-aided detection: a population-based study of medicare enrollees. Ann Intern Med 2013; 158:580-7. [PMID: 23588746 PMCID: PMC3772716 DOI: 10.7326/0003-4819-158-8-201304160-00002] [Citation(s) in RCA: 42] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 11/22/2022] Open
Abstract
BACKGROUND Computer-aided detection (CAD) has rapidly diffused into screening mammography practice despite limited and conflicting data on its clinical effect. OBJECTIVE To determine associations between CAD use during screening mammography and the incidence of ductal carcinoma in situ (DCIS) and invasive breast cancer, invasive cancer stage, and diagnostic testing. DESIGN Retrospective cohort study. SETTING Medicare program. PARTICIPANTS Women aged 67 to 89 years having screening mammography between 2001 and 2006 in U.S. SEER (Surveillance, Epidemiology and End Results) regions (409 459 mammograms from 163 099 women). MEASUREMENTS Incident DCIS and invasive breast cancer within 1 year after mammography, invasive cancer stage, and diagnostic testing within 90 days after screening among women without breast cancer. RESULTS From 2001 to 2006, CAD prevalence increased from 3.6% to 60.5%. Use of CAD was associated with greater DCIS incidence (adjusted odds ratio [OR], 1.17 [95% CI, 1.11 to 1.23]) but no difference in invasive breast cancer incidence (adjusted OR, 1.00 [CI, 0.97 to 1.03]). Among women with invasive cancer, CAD was associated with greater likelihood of stage I to II versus III to IV cancer (adjusted OR, 1.27 [CI, 1.14 to 1.41]). In women without breast cancer, CAD was associated with increased odds of diagnostic mammography (adjusted OR, 1.28 [CI, 1.27 to 1.29]), breast ultrasonography (adjusted OR, 1.07 [CI, 1.06 to 1.09]), and breast biopsy (adjusted OR, 1.10 [CI, 1.08 to 1.12]). LIMITATION Short follow-up for cancer stage, potential unmeasured confounding, and uncertain generalizability to younger women. CONCLUSION Use of CAD during screening mammography among Medicare enrollees is associated with increased DCIS incidence, the diagnosis of invasive breast cancer at earlier stages, and increased diagnostic testing among women without breast cancer. PRIMARY FUNDING SOURCE Center for Healthcare Policy and Research, University of California, Davis.
Collapse
Affiliation(s)
- Joshua J Fenton
- University of California, Davis, Department of Family and Community Medicine, 4860 Y Street, Suite 2300, Sacramento, CA 95817, USA.
| | | | | | | | | | | | | |
Collapse
|