1
|
Weber S, Wyszynski M, Godefroid M, Plattfaut R, Niehaves B. How do medical professionals make sense (or not) of AI? A social-media-based computational grounded theory study and an online survey. Comput Struct Biotechnol J 2024; 24:146-159. [PMID: 38434249 PMCID: PMC10904922 DOI: 10.1016/j.csbj.2024.02.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2023] [Revised: 02/14/2024] [Accepted: 02/14/2024] [Indexed: 03/05/2024] Open
Abstract
To investigate opinions and attitudes of medical professionals towards adopting AI-enabled healthcare technologies in their daily business, we used a mixed-methods approach. Study 1 employed a qualitative computational grounded theory approach analyzing 181 Reddit threads in the several subreddits of r/medicine. By utilizing an unsupervised machine learning clustering method, we identified three key themes: (1) consequences of AI, (2) physician-AI relationship, and (3) a proposed way forward. In particular Reddit posts related to the first two themes indicated that the medical professionals' fear of being replaced by AI and skepticism toward AI played a major role in the argumentations. Moreover, the results suggest that this fear is driven by little or moderate knowledge about AI. Posts related to the third theme focused on factual discussions about how AI and medicine have to be designed to become broadly adopted in health care. Study 2 quantitatively examined the relationship between the fear of AI, knowledge about AI, and medical professionals' intention to use AI-enabled technologies in more detail. Results based on a sample of 223 medical professionals who participated in the online survey revealed that the intention to use AI technologies increases with increasing knowledge about AI and that this effect is moderated by the fear of being replaced by AI.
Collapse
Affiliation(s)
- Sebastian Weber
- University of Bremen, Digital Public, Bibliothekstr. 1, 28359 Bremen, Germany
| | - Marc Wyszynski
- University of Bremen, Digital Public, Bibliothekstr. 1, 28359 Bremen, Germany
| | - Marie Godefroid
- University of Siegen, Information Systems, Kohlbettstr. 15, 57072 Siegen, Germany
| | - Ralf Plattfaut
- University of Duisburg-Essen, Information Systems and Transformation Management, Universitätsstr. 9, 45141 Essen, Germany
| | - Bjoern Niehaves
- University of Bremen, Digital Public, Bibliothekstr. 1, 28359 Bremen, Germany
| |
Collapse
|
2
|
Garfinkle J. Automation to approximate the Gestalt: applying machine learning to the general movement assessment. Pediatr Res 2024:10.1038/s41390-024-03558-w. [PMID: 39251880 DOI: 10.1038/s41390-024-03558-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/17/2024] [Accepted: 08/24/2024] [Indexed: 09/11/2024]
Affiliation(s)
- Jarred Garfinkle
- Division of Neonatology, Department of Pediatrics, McGill University, Montreal Children's Hospital, Montreal, QC, Canada.
| |
Collapse
|
3
|
Makkar J, Flores J, Matich M, Duong TT, Thompson SM, Du Y, Busch I, Phan QM, Wang Q, Delevich K, Broughton-Neiswanger L, Driskell IM, Driskell RR. Deep Hair Phenomics: Implications in Endocrinology, Development, and Aging. J Invest Dermatol 2024:S0022-202X(24)02079-7. [PMID: 39236901 DOI: 10.1016/j.jid.2024.08.014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2024] [Revised: 08/08/2024] [Accepted: 08/11/2024] [Indexed: 09/07/2024]
Abstract
Hair quality is an important indicator of health in humans and other animals. Current approaches to assess hair quality are generally non-quantitative or are low throughput due to technical limitations of 'splitting hairs'. We developed a deep learning-based computer vision approach for the high throughput quantification of individual hair fibers at a high resolution. Our innovative computer vision tool can distinguish and extract overlapping fibers for quantification of multivariate features including length, width, and color to generate single-hair phenomes (shPhenome) of diverse conditions across the lifespan of mice. Using our tool, we explored the effects of hormone signaling, genetic modifications, and aging on hair follicle output. Our analyses revealed hair phenotypes resultant of endocrinological, developmental, and aging-related alterations in the fur coats of mice. These results demonstrate the efficacy of our deep hair phenomics tool for characterizing factors that modulate the hair follicle and developing new diagnostic methods for detecting disease through the hair fiber. Finally, we have generated a searchable, interactive web tool for the exploration of our hair fiber data at skinregeneration.org.
Collapse
Affiliation(s)
- Jasson Makkar
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA
| | - Jorge Flores
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA
| | - Mason Matich
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA
| | - Tommy T Duong
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA
| | - Sean M Thompson
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA
| | - Yiqing Du
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA
| | - Isabelle Busch
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA
| | - Quan M Phan
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA
| | - Qing Wang
- Department of Integrative Physiology and Neuroscience, Washington State University, Pullman, WA, 99164, USA
| | - Kristen Delevich
- Center for Reproductive Biology, Washington State University, Pullman, WA, 99164, USA; Department of Integrative Physiology and Neuroscience, Washington State University, Pullman, WA, 99164, USA
| | - Liam Broughton-Neiswanger
- Washington Animal Disease Diagnostic Laboratory, Washington State University, Pullman, Washington, 99164, USA
| | - Iwona M Driskell
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA
| | - Ryan R Driskell
- School of Molecular Biosciences, Washington State University, 100 Dairy Road, BLS 233, Pullman, Washington, 99164, USA; Center for Reproductive Biology, Washington State University, Pullman, WA, 99164, USA.
| |
Collapse
|
4
|
Sams CM, Fanous AH, Daneshjou R. Human-Artificial Intelligence Interaction Research Is Crucial for Medical Artificial Intelligence Implementation. J Invest Dermatol 2024:S0022-202X(24)01976-6. [PMID: 39230537 DOI: 10.1016/j.jid.2024.07.014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2024] [Revised: 07/15/2024] [Accepted: 07/18/2024] [Indexed: 09/05/2024]
Affiliation(s)
- Clarence M Sams
- Department of Dermatology, Stanford School of Medicine, Stanford, California, USA; Department of Biomedical Data Science, Stanford School of Medicine, Stanford, California, USA; College of Medicine, Charles R. Drew University of Medicine and Science, Los Angeles, California, USA
| | - Aaron H Fanous
- Department of Biomedical Data Science, Stanford School of Medicine, Stanford, California, USA
| | - Roxana Daneshjou
- Department of Dermatology, Stanford School of Medicine, Stanford, California, USA; Department of Biomedical Data Science, Stanford School of Medicine, Stanford, California, USA.
| |
Collapse
|
5
|
Kaur J, Kaur P. A systematic literature analysis of multi-organ cancer diagnosis using deep learning techniques. Comput Biol Med 2024; 179:108910. [PMID: 39032244 DOI: 10.1016/j.compbiomed.2024.108910] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2024] [Revised: 07/14/2024] [Accepted: 07/15/2024] [Indexed: 07/23/2024]
Abstract
Cancer is becoming the most toxic ailment identified among individuals worldwide. The mortality rate has been increasing rapidly every year, which causes progression in the various diagnostic technologies to handle this illness. The manual procedure for segmentation and classification with a large set of data modalities can be a challenging task. Therefore, a crucial requirement is to significantly develop the computer-assisted diagnostic system intended for the initial cancer identification. This article offers a systematic review of Deep Learning approaches using various image modalities to detect multi-organ cancers from 2012 to 2023. It emphasizes the detection of five supreme predominant tumors, i.e., breast, brain, lung, skin, and liver. Extensive review has been carried out by collecting research and conference articles and book chapters from reputed international databases, i.e., Springer Link, IEEE Xplore, Science Direct, PubMed, and Wiley that fulfill the criteria for quality evaluation. This systematic review summarizes the overview of convolutional neural network model architectures and datasets used for identifying and classifying the diverse categories of cancer. This study accomplishes an inclusive idea of ensemble deep learning models that have achieved better evaluation results for classifying the different images into cancer or healthy cases. This paper will provide a broad understanding to the research scientists within the domain of medical imaging procedures of which deep learning technique perform best over which type of dataset, extraction of features, different confrontations, and their anticipated solutions for the complex problems. Lastly, some challenges and issues which control the health emergency have been discussed.
Collapse
Affiliation(s)
- Jaspreet Kaur
- Department of Computer Engineering & Technology, Guru Nanak Dev University, Amritsar, Punjab, India.
| | - Prabhpreet Kaur
- Department of Computer Engineering & Technology, Guru Nanak Dev University, Amritsar, Punjab, India.
| |
Collapse
|
6
|
Herrmann M. What Are Patients Doing in the Loop? Patients as Fellow-Workers in the Everyday Use of Medical AI. THE AMERICAN JOURNAL OF BIOETHICS : AJOB 2024; 24:91-93. [PMID: 39225997 DOI: 10.1080/15265161.2024.2377122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/04/2024]
Affiliation(s)
- Markus Herrmann
- National Center for Tumor Diseases (NCT) Heidelberg, a partnership between DKFZ and Heidelberg University Hospital, German Cancer Research Center (DKFZ), Heidelberg University, Heidelberg University Hospital
| |
Collapse
|
7
|
Mellers BA, McCoy JP, Lu L, Tetlock PE. Human and Algorithmic Predictions in Geopolitical Forecasting: Quantifying Uncertainty in Hard-to-Quantify Domains. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2024; 19:711-721. [PMID: 37642169 PMCID: PMC11373164 DOI: 10.1177/17456916231185339] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/31/2023]
Abstract
Research on clinical versus statistical prediction has demonstrated that algorithms make more accurate predictions than humans in many domains. Geopolitical forecasting is an algorithm-unfriendly domain, with hard-to-quantify data and elusive reference classes that make predictive model-building difficult. Furthermore, the stakes can be high, with missed forecasts leading to mass-casualty consequences. For these reasons, geopolitical forecasting is typically done by humans, even though algorithms play important roles. They are essential as aggregators of crowd wisdom, as frameworks to partition human forecasting variance, and as inputs to hybrid forecasting models. Algorithms are extremely important in this domain. We doubt that humans will relinquish control to algorithms anytime soon-nor do we think they should. However, the accuracy of forecasts will greatly improve if humans are aided by algorithms.
Collapse
Affiliation(s)
| | - John P McCoy
- Department of Marketing, University of Pennsylvania
| | - Louise Lu
- Department of Marketing, Stanford Business School, Stanford University
| | - Philip E Tetlock
- Management Department of Wharton Business School, University of Pennsylvania
| |
Collapse
|
8
|
Winkler JK, Kommoss KS, Vollmer AS, Blum A, Stolz W, Kränke T, Hofmann-Wellenhof R, Enk A, Toberer F, Haenssle HA. Computerizing the first step of the two-step algorithm in dermoscopy: A convolutional neural network for differentiating melanocytic from non-melanocytic skin lesions. Eur J Cancer 2024; 210:114297. [PMID: 39217816 DOI: 10.1016/j.ejca.2024.114297] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2024] [Revised: 08/14/2024] [Accepted: 08/15/2024] [Indexed: 09/04/2024]
Abstract
IMPORTANCE Convolutional neural networks (CNN) have shown performance equal to trained dermatologists in differentiating benign from malignant skin lesions. To improve clinicians' management decisions, additional classifications into diagnostic categories might be helpful. METHODS A convenience sample of 100 pigmented/non-pigmented skin lesions was used for a cross-sectional two-level reader study including 96 dermatologists (level I: dermoscopy only; level II: clinical close-up images, dermoscopy, and textual information). Dermoscopic images were classified by a binary CNN trained to differentiate melanocytic from non-melanocytic lesions (FotoFinder Systems, Bad Birnbach, Germany). Primary endpoint was the accuracy of the CNN's classification in comparison with dermatologists reviewing level-II information. Secondary endpoints included dermatologists' accuracies according to their level of experience and the CNN's area under the curve (AUC) of receiver operating characteristics (ROC). RESULTS The CNN revealed an accuracy and ROC AUC with corresponding 95 % confidence intervals (CI) of 91.0 % (83.8 % to 95.2 %) and 0.981 (0.962 to 1). In level I, dermatologists showed a mean accuracy of 83.7 % (82.5 % to 84.8 %). With level II information, the accuracy improved to 87.8 % (86.7 % to 88.9 %; p < 0.001). When comparing accuracies of CNN and dermatologists in level II, the CNN's accuracy was higher (91.0 % versus 87.8 %, p < 0.001). For experts with level II information results were on par with the CNN (91.0 % versus 90.4 %, p = 0.368). CONCLUSIONS The tested CNN accurately differentiated melanocytic from non-melanocytic skin lesions and outperformed dermatologists. The CNN may support clinicians and could be used in an ensemble approach combined with other CNN models.
Collapse
Affiliation(s)
- Julia K Winkler
- Department of Dermatology, University of Heidelberg, Heidelberg, Germany.
| | | | | | - Andreas Blum
- Public, Private and Teaching Practice, Konstanz, Germany
| | - Wilhelm Stolz
- Department of Dermatology, Allergology and Environmental Medicine II, Hospital Thalkirchner Street, Munich, Germany
| | - T Kränke
- Department of Dermatology and Venerology, Medical University of Graz, Graz, Austria
| | - R Hofmann-Wellenhof
- Department of Dermatology and Venerology, Medical University of Graz, Graz, Austria
| | - Alexander Enk
- Department of Dermatology, University of Heidelberg, Heidelberg, Germany
| | - Ferdinand Toberer
- Department of Dermatology, University of Heidelberg, Heidelberg, Germany
| | - Holger A Haenssle
- Department of Dermatology, University of Heidelberg, Heidelberg, Germany
| |
Collapse
|
9
|
Banharak S, Threeranut A, Metprommarat A. Effects of health education applying role play on decision making in acute myocardial infarction situation among older adults in community. Sci Rep 2024; 14:19627. [PMID: 39179621 PMCID: PMC11344021 DOI: 10.1038/s41598-024-70747-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/20/2024] [Accepted: 08/20/2024] [Indexed: 08/26/2024] Open
Abstract
Acute myocardial infarction is a silent killer for people worldwide, especially older adults who often experience atypical symptoms, causing late decision-making and a high mortality rate. The unrecognition of atypical symptoms, unconcerned about their risk, and not knowing how to deal with this critical situation are the barriers to a quick decision to visit the emergency department and delaying treatment, resulting in serious adverse outcomes. Therefore, specific and effective health education among older adults is needed. This double-blinded randomized controlled trial explored the effectiveness of health education by applying a role-play promoting decision-making ability program when expecting acute myocardial infarction occurrence among community-dwelling older adults. The participants were 96 community-dwelling older adults in central northeastern Thailand. We collected data between November 2021 and April 2022. The multi-stage sampling was applied to include participants. The intervention was the role-play promoting decision-making ability program and home visit. Outcomes were measured a week before attending and after finishing the intervention. T-tests, Mann-Whitney U test, Chi-square, and Wilcoxon Signed Rank test compared the outcomes between and within groups. Moreover, adjusted analysis was also demonstrated. Results revealed that participants who attended the program improved their knowledge, belief, and decision-making; only perceived susceptibility did not show improvement. Moreover, after demonstrating an adjusted analysis, the program participants had better knowledge about symptoms, perceived benefits, barriers, self-regulation, possible calling 1669, and first action. In conclusion, a role-play promoting decision-making ability program can promote knowledge, belief, and decision-making when expecting acute myocardial infarction occurrence among community-dwelling older adults. This study proved that role-play is one strategy to promote the program's effectiveness by inducing attention before giving older adults health information. Nurses and other healthcare professionals can implement this program as part of standard practice.Clinical Trial Registration Number: TCTR20210928004 on 28/09/2021.
Collapse
Affiliation(s)
- Samoraphop Banharak
- Department of Gerontological Nursing, Faculty of Nursing, Khon Kaen University, Khon Kaen, Thailand.
| | - Ampornpan Threeranut
- Department of Adult Nursing, Faculty of Nursing, Khon Kaen University, Khon Kaen, Thailand
| | | |
Collapse
|
10
|
Indolfi C, Klain A, Dinardo G, Decimo F, Miraglia del Giudice M. Artificial intelligence in the transition of allergy: a valuable tool from childhood to adulthood. Front Med (Lausanne) 2024; 11:1469161. [PMID: 39219791 PMCID: PMC11363185 DOI: 10.3389/fmed.2024.1469161] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2024] [Accepted: 08/01/2024] [Indexed: 09/04/2024] Open
Affiliation(s)
| | - Angela Klain
- Department of Woman, Child and General and Specialized Surgery, University of Campania 'Luigi Vanvitelli', Naples, Italy
| | - Giulio Dinardo
- Department of Woman, Child and General and Specialized Surgery, University of Campania 'Luigi Vanvitelli', Naples, Italy
| | | | | |
Collapse
|
11
|
Kurtansky NR, D'Alessandro BM, Gillis MC, Betz-Stablein B, Cerminara SE, Garcia R, Girundi MA, Goessinger EV, Gottfrois P, Guitera P, Halpern AC, Jakrot V, Kittler H, Kose K, Liopyris K, Malvehy J, Mar VJ, Martin LK, Mathew T, Maul LV, Mothershaw A, Mueller AM, Mueller C, Navarini AA, Rajeswaran T, Rajeswaran V, Saha A, Sashindranath M, Serra-García L, Soyer HP, Theocharis G, Vos A, Weber J, Rotemberg V. The SLICE-3D dataset: 400,000 skin lesion image crops extracted from 3D TBP for skin cancer detection. Sci Data 2024; 11:884. [PMID: 39143096 PMCID: PMC11324883 DOI: 10.1038/s41597-024-03743-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2024] [Accepted: 08/05/2024] [Indexed: 08/16/2024] Open
Abstract
AI image classification algorithms have shown promising results when applied to skin cancer detection. Most public skin cancer image datasets are comprised of dermoscopic photos and are limited by selection bias, lack of standardization, and lend themselves to development of algorithms that can only be used by skilled clinicians. The SLICE-3D ("Skin Lesion Image Crops Extracted from 3D TBP") dataset described here addresses those concerns and contains images of over 400,000 distinct skin lesions from seven dermatologic centers from around the world. De-identified images were systematically extracted from sensitive 3D Total Body Photographs and are comparable in optical resolution to smartphone images. Algorithms trained on lower quality images could improve clinical workflows and detect skin cancers earlier if deployed in primary care or non-clinical settings, where photos are captured by non-expert physicians or patients. Such a tool could prompt individuals to visit a specialized dermatologist. This dataset circumvents many inherent limitations of prior datasets and may be used to build upon previous applications of skin imaging for cancer detection.
Collapse
Affiliation(s)
- Nicholas R Kurtansky
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, New York, USA.
| | | | - Maura C Gillis
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, New York, USA
| | - Brigid Betz-Stablein
- Frazer Institute, The University of Queensland, Dermatology Research Centre, Brisbane, Queensland, Australia
| | - Sara E Cerminara
- Department of Dermatology, University Hospital of Basel, Basel, Switzerland
| | - Rafael Garcia
- Computer Vision and Robotics Institute, University of Girona, Girona, Spain
| | | | | | - Philippe Gottfrois
- Department of Dermatology, University Hospital of Basel, Basel, Switzerland
| | - Pascale Guitera
- Melanoma Institute Australia, Sydney, Australia
- Faculty of Medicine and Health, University of Sydney, Sydney, Australia
| | - Allan C Halpern
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, New York, USA
| | | | - Harald Kittler
- ViDIR Group, Department of Dermatology, Medical University of Vienna, Vienna, Austria
| | - Kivanc Kose
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, New York, USA
| | | | - Josep Malvehy
- Dermatology Department, Hospital Clínic Barcelona, Universitat de Barcelona, IDIBAPS, Barcelona, Spain
- Centro de Investigación Biomédica en Red de Enfermedades Raras (CIBER ER), Instituto de Salud Carlos III, Barcelona, Spain
| | - Victoria J Mar
- School of Public Health and Preventive Medicine, Monash University, Melbourne, VIC, Australia
- Victorian Melanoma Service, Alfred Hospital, 55 Commercial Road, Melbourne, VIC, 3004, Australia
| | - Linda K Martin
- Melanoma Institute Australia, Sydney, Australia
- Faculty of Medicine and Health, University of Sydney, Sydney, Australia
- School of Clinical Medicine, Faculty of Medicine & Health, University of New South Wales, Sydney, Australia
| | | | - Lara Valeska Maul
- Department of Dermatology, University Hospital of Zurich, Zurich, Switzerland
| | - Adam Mothershaw
- Frazer Institute, The University of Queensland, Dermatology Research Centre, Brisbane, Queensland, Australia
| | - Alina M Mueller
- Department of Dermatology, University Hospital of Basel, Basel, Switzerland
| | - Christoph Mueller
- ViDIR Group, Department of Dermatology, Medical University of Vienna, Vienna, Austria
| | | | | | | | - Anup Saha
- Computer Vision and Robotics Institute, University of Girona, Girona, Spain
| | - Maithili Sashindranath
- School of Public Health and Preventive Medicine, Monash University, Melbourne, VIC, Australia
| | | | - H Peter Soyer
- Frazer Institute, The University of Queensland, Dermatology Research Centre, Brisbane, Queensland, Australia
| | | | - Ayesha Vos
- Victorian Melanoma Service, Alfred Hospital, 55 Commercial Road, Melbourne, VIC, 3004, Australia
| | - Jochen Weber
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, New York, USA
| | - Veronica Rotemberg
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, New York, USA
| |
Collapse
|
12
|
Monnier J, Foahom Gouabou AC, Serdi M, Collenne J, Iguernaissi R, Richard MA, Gaudy-Marqueste C, Damoiseaux JL, Grob JJ, Merad D. Automated melanoma detection. An algorithm inspired from human intelligence characterizing disordered pattern of melanocytic lesions improving a convolutional neural network. J Am Acad Dermatol 2024; 91:350-352. [PMID: 38614250 DOI: 10.1016/j.jaad.2024.02.063] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2023] [Revised: 01/23/2024] [Accepted: 02/27/2024] [Indexed: 04/15/2024]
Affiliation(s)
- Jilliana Monnier
- Aix-Marseille Univ, Université de Marseille, CNRS UMR 7020, Laboratoire d'Informatique et Systèmes, Marseille, France; Centre de Recherche en Cancérologie de Marseille, Inserm UMR1068, CNRS UMR7258, Aix-Marseille University, Marseille, France; Dermatology and Skin cancer department, La Timone Hospital, Assistance Publique-Hôpitaux de Marseille, Aix-Marseille University, Marseille, France.
| | - Arthur Cartel Foahom Gouabou
- Aix-Marseille Univ, Université de Marseille, CNRS UMR 7020, Laboratoire d'Informatique et Systèmes, Marseille, France
| | - Meryem Serdi
- Aix-Marseille Univ, Université de Marseille, CNRS UMR 7020, Laboratoire d'Informatique et Systèmes, Marseille, France
| | - Jules Collenne
- Aix-Marseille Univ, Université de Marseille, CNRS UMR 7020, Laboratoire d'Informatique et Systèmes, Marseille, France
| | - Rabah Iguernaissi
- Aix-Marseille Univ, Université de Marseille, CNRS UMR 7020, Laboratoire d'Informatique et Systèmes, Marseille, France
| | - Marie-Aleth Richard
- Dermatology and Skin cancer department, La Timone Hospital, Assistance Publique-Hôpitaux de Marseille, Aix-Marseille University, Marseille, France
| | - Caroline Gaudy-Marqueste
- Centre de Recherche en Cancérologie de Marseille, Inserm UMR1068, CNRS UMR7258, Aix-Marseille University, Marseille, France; Dermatology and Skin cancer department, La Timone Hospital, Assistance Publique-Hôpitaux de Marseille, Aix-Marseille University, Marseille, France
| | - Jean-Luc Damoiseaux
- Aix-Marseille Univ, Université de Marseille, CNRS UMR 7020, Laboratoire d'Informatique et Systèmes, Marseille, France
| | - Jean-Jacques Grob
- Centre de Recherche en Cancérologie de Marseille, Inserm UMR1068, CNRS UMR7258, Aix-Marseille University, Marseille, France; Dermatology and Skin cancer department, La Timone Hospital, Assistance Publique-Hôpitaux de Marseille, Aix-Marseille University, Marseille, France
| | - Djamal Merad
- Aix-Marseille Univ, Université de Marseille, CNRS UMR 7020, Laboratoire d'Informatique et Systèmes, Marseille, France
| |
Collapse
|
13
|
Tognetti L, Miracapillo C, Leonardelli S, Luschi A, Iadanza E, Cevenini G, Rubegni P, Cartocci A. Deep Learning Techniques for the Dermoscopic Differential Diagnosis of Benign/Malignant Melanocytic Skin Lesions: From the Past to the Present. Bioengineering (Basel) 2024; 11:758. [PMID: 39199716 PMCID: PMC11351129 DOI: 10.3390/bioengineering11080758] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2024] [Revised: 07/15/2024] [Accepted: 07/17/2024] [Indexed: 09/01/2024] Open
Abstract
There has been growing scientific interest in the research field of deep learning techniques applied to skin cancer diagnosis in the last decade. Though encouraging data have been globally reported, several discrepancies have been observed in terms of study methodology, result presentations and validation in clinical settings. The present review aimed to screen the scientific literature on the application of DL techniques to dermoscopic melanoma/nevi differential diagnosis and extrapolate those original studies adequately by reporting on a DL model, comparing them among clinicians and/or another DL architecture. The second aim was to examine those studies together according to a standard set of statistical measures, and the third was to provide dermatologists with a comprehensive explanation and definition of the most used artificial intelligence (AI) terms to better/further understand the scientific literature on this topic and, in parallel, to be updated on the newest applications in the medical dermatologic field, along with a historical perspective. After screening nearly 2000 records, a subset of 54 was selected. Comparing the 20 studies reporting on convolutional neural network (CNN)/deep convolutional neural network (DCNN) models, we have a scenario of highly performant DL algorithms, especially in terms of low false positive results, with average values of accuracy (83.99%), sensitivity (77.74%), and specificity (80.61%). Looking at the comparison with diagnoses by clinicians (13 studies), the main difference relies on the specificity values, with a +15.63% increase for the CNN/DCNN models (average specificity of 84.87%) compared to humans (average specificity of 64.24%) with a 14,85% gap in average accuracy; the sensitivity values were comparable (79.77% for DL and 79.78% for humans). To obtain higher diagnostic accuracy and feasibility in clinical practice, rather than in experimental retrospective settings, future DL models should be based on a large dataset integrating dermoscopic images with relevant clinical and anamnestic data that is prospectively tested and adequately compared with physicians.
Collapse
Affiliation(s)
- Linda Tognetti
- Dermatology Unit, Deparment of Medical, Surgical and Neurosciences, University of Siena, Viale Bracci 16, 53100 Siena, Italy (P.R.)
| | - Chiara Miracapillo
- Dermatology Unit, Deparment of Medical, Surgical and Neurosciences, University of Siena, Viale Bracci 16, 53100 Siena, Italy (P.R.)
| | - Simone Leonardelli
- Dermatology Unit, Deparment of Medical, Surgical and Neurosciences, University of Siena, Viale Bracci 16, 53100 Siena, Italy (P.R.)
| | - Alessio Luschi
- Bioengineering and Biomedical Data Science Lab, Department of Medical Biotechnologies, University of Siena, 53100 Siena, Italy (E.I.)
| | - Ernesto Iadanza
- Bioengineering and Biomedical Data Science Lab, Department of Medical Biotechnologies, University of Siena, 53100 Siena, Italy (E.I.)
| | - Gabriele Cevenini
- Bioengineering and Biomedical Data Science Lab, Department of Medical Biotechnologies, University of Siena, 53100 Siena, Italy (E.I.)
| | - Pietro Rubegni
- Dermatology Unit, Deparment of Medical, Surgical and Neurosciences, University of Siena, Viale Bracci 16, 53100 Siena, Italy (P.R.)
| | - Alessandra Cartocci
- Dermatology Unit, Deparment of Medical, Surgical and Neurosciences, University of Siena, Viale Bracci 16, 53100 Siena, Italy (P.R.)
- Bioengineering and Biomedical Data Science Lab, Department of Medical Biotechnologies, University of Siena, 53100 Siena, Italy (E.I.)
| |
Collapse
|
14
|
Gomez C, Smith BL, Zayas A, Unberath M, Canares T. Explainable AI decision support improves accuracy during telehealth strep throat screening. COMMUNICATIONS MEDICINE 2024; 4:149. [PMID: 39048726 PMCID: PMC11269612 DOI: 10.1038/s43856-024-00568-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2023] [Accepted: 07/04/2024] [Indexed: 07/27/2024] Open
Abstract
BACKGROUND Artificial intelligence-based (AI) clinical decision support systems (CDSS) using unconventional data, like smartphone-acquired images, promise transformational opportunities for telehealth; including remote diagnosis. Although such solutions' potential remains largely untapped, providers' trust and understanding are vital for effective adoption. This study examines how different human-AI interaction paradigms affect clinicians' responses to an emerging AI CDSS for streptococcal pharyngitis (strep throat) detection from smartphone throat images. METHODS In a randomized experiment, we tested explainable AI strategies using three AI-based CDSS prototypes for strep throat prediction. Participants received clinical vignettes via an online survey to predict the disease state and offer clinical recommendations. The first set included a validated CDSS prediction (Modified Centor Score) and the second introduced an explainable AI prototype randomly. We used linear models to assess explainable AI's effect on clinicians' accuracy, confirmatory testing rates, and perceived trust and understanding of the CDSS. RESULTS The study, involving 121 telehealth providers, shows that compared to using the Centor Score, AI-based CDSS can improve clinicians' predictions. Despite higher agreement with AI, participants report lower trust in its advice than in the Centor Score, leading to more requests for in-person confirmatory testing. CONCLUSIONS Effectively integrating AI is crucial in the telehealth-based diagnosis of infectious diseases, given the implications of antibiotic over-prescriptions. We demonstrate that AI-based CDSS can improve the accuracy of remote strep throat screening yet underscores the necessity to enhance human-machine collaboration, particularly in trust and intelligibility. This ensures providers and patients can capitalize on AI interventions and smartphones for virtual healthcare.
Collapse
Affiliation(s)
- Catalina Gomez
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
| | | | - Alisa Zayas
- Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Mathias Unberath
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA.
- Johns Hopkins University School of Medicine, Baltimore, MD, USA.
| | - Therese Canares
- Division of Pediatric Emergency Medicine, Johns Hopkins University School of Medicine, Baltimore, MD, USA
| |
Collapse
|
15
|
Hager P, Jungmann F, Holland R, Bhagat K, Hubrecht I, Knauer M, Vielhauer J, Makowski M, Braren R, Kaissis G, Rueckert D. Evaluation and mitigation of the limitations of large language models in clinical decision-making. Nat Med 2024:10.1038/s41591-024-03097-1. [PMID: 38965432 DOI: 10.1038/s41591-024-03097-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2024] [Accepted: 05/29/2024] [Indexed: 07/06/2024]
Abstract
Clinical decision-making is one of the most impactful parts of a physician's responsibilities and stands to benefit greatly from artificial intelligence solutions and large language models (LLMs) in particular. However, while LLMs have achieved excellent performance on medical licensing exams, these tests fail to assess many skills necessary for deployment in a realistic clinical decision-making environment, including gathering information, adhering to guidelines, and integrating into clinical workflows. Here we have created a curated dataset based on the Medical Information Mart for Intensive Care database spanning 2,400 real patient cases and four common abdominal pathologies as well as a framework to simulate a realistic clinical setting. We show that current state-of-the-art LLMs do not accurately diagnose patients across all pathologies (performing significantly worse than physicians), follow neither diagnostic nor treatment guidelines, and cannot interpret laboratory results, thus posing a serious risk to the health of patients. Furthermore, we move beyond diagnostic accuracy and demonstrate that they cannot be easily integrated into existing workflows because they often fail to follow instructions and are sensitive to both the quantity and order of information. Overall, our analysis reveals that LLMs are currently not ready for autonomous clinical decision-making while providing a dataset and framework to guide future studies.
Collapse
Affiliation(s)
- Paul Hager
- Institute for AI and Informatics, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany.
- Institute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany.
| | - Friederike Jungmann
- Institute for AI and Informatics, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
- Institute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
| | | | - Kunal Bhagat
- Department of Medicine, ChristianaCare Health System, Wilmington, DE, USA
| | - Inga Hubrecht
- Department of Medicine III, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
| | - Manuel Knauer
- Department of Medicine III, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
| | - Jakob Vielhauer
- Department of Medicine II, University Hospital of the Ludwig Maximilian University of Munich, Munich, Germany
| | - Marcus Makowski
- Institute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
| | - Rickmer Braren
- Institute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
| | - Georgios Kaissis
- Institute for AI and Informatics, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
- Institute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
- Department of Computing, Imperial College, London, UK
- Reliable AI Group, Institute for Machine Learning in Biomedical Imaging, Helmholtz Munich, Munich, Germany
| | - Daniel Rueckert
- Institute for AI and Informatics, Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
- Department of Computing, Imperial College, London, UK
| |
Collapse
|
16
|
Duggan NM, Jin M, Duran Mendicuti MA, Hallisey S, Bernier D, Selame LA, Asgari-Targhi A, Fischetti CE, Lucassen R, Samir AE, Duhaime E, Kapur T, Goldsmith AJ. Gamified Crowdsourcing as a Novel Approach to Lung Ultrasound Data Set Labeling: Prospective Analysis. J Med Internet Res 2024; 26:e51397. [PMID: 38963923 PMCID: PMC11258523 DOI: 10.2196/51397] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2023] [Revised: 10/04/2023] [Accepted: 04/10/2024] [Indexed: 07/06/2024] Open
Abstract
BACKGROUND Machine learning (ML) models can yield faster and more accurate medical diagnoses; however, developing ML models is limited by a lack of high-quality labeled training data. Crowdsourced labeling is a potential solution but can be constrained by concerns about label quality. OBJECTIVE This study aims to examine whether a gamified crowdsourcing platform with continuous performance assessment, user feedback, and performance-based incentives could produce expert-quality labels on medical imaging data. METHODS In this diagnostic comparison study, 2384 lung ultrasound clips were retrospectively collected from 203 emergency department patients. A total of 6 lung ultrasound experts classified 393 of these clips as having no B-lines, one or more discrete B-lines, or confluent B-lines to create 2 sets of reference standard data sets (195 training clips and 198 test clips). Sets were respectively used to (1) train users on a gamified crowdsourcing platform and (2) compare the concordance of the resulting crowd labels to the concordance of individual experts to reference standards. Crowd opinions were sourced from DiagnosUs (Centaur Labs) iOS app users over 8 days, filtered based on past performance, aggregated using majority rule, and analyzed for label concordance compared with a hold-out test set of expert-labeled clips. The primary outcome was comparing the labeling concordance of collated crowd opinions to trained experts in classifying B-lines on lung ultrasound clips. RESULTS Our clinical data set included patients with a mean age of 60.0 (SD 19.0) years; 105 (51.7%) patients were female and 114 (56.1%) patients were White. Over the 195 training clips, the expert-consensus label distribution was 114 (58%) no B-lines, 56 (29%) discrete B-lines, and 25 (13%) confluent B-lines. Over the 198 test clips, expert-consensus label distribution was 138 (70%) no B-lines, 36 (18%) discrete B-lines, and 24 (12%) confluent B-lines. In total, 99,238 opinions were collected from 426 unique users. On a test set of 198 clips, the mean labeling concordance of individual experts relative to the reference standard was 85.0% (SE 2.0), compared with 87.9% crowdsourced label concordance (P=.15). When individual experts' opinions were compared with reference standard labels created by majority vote excluding their own opinion, crowd concordance was higher than the mean concordance of individual experts to reference standards (87.4% vs 80.8%, SE 1.6 for expert concordance; P<.001). Clips with discrete B-lines had the most disagreement from both the crowd consensus and individual experts with the expert consensus. Using randomly sampled subsets of crowd opinions, 7 quality-filtered opinions were sufficient to achieve near the maximum crowd concordance. CONCLUSIONS Crowdsourced labels for B-line classification on lung ultrasound clips via a gamified approach achieved expert-level accuracy. This suggests a strategic role for gamified crowdsourcing in efficiently generating labeled image data sets for training ML systems.
Collapse
Affiliation(s)
- Nicole M Duggan
- Department of Emergency Medicine, Brigham and Women's Hospital, Harvard Medical School, Boston, MA, United States
| | - Mike Jin
- Department of Emergency Medicine, Brigham and Women's Hospital, Harvard Medical School, Boston, MA, United States
- Centaur Labs, Boston, MA, United States
| | | | - Stephen Hallisey
- Department of Emergency Medicine, Brigham and Women's Hospital, Harvard Medical School, Boston, MA, United States
| | - Denie Bernier
- Department of Emergency Medicine, Brigham and Women's Hospital, Harvard Medical School, Boston, MA, United States
| | - Lauren A Selame
- Department of Emergency Medicine, Brigham and Women's Hospital, Harvard Medical School, Boston, MA, United States
| | - Ameneh Asgari-Targhi
- Department of Radiology, Brigham and Women's Hospital, Harvard Medical School, Boston, MA, United States
| | - Chanel E Fischetti
- Department of Emergency Medicine, Brigham and Women's Hospital, Harvard Medical School, Boston, MA, United States
| | - Ruben Lucassen
- Department of Biomedical Engineering, Eindhoven University of Technology, Eindhoven, Netherlands
| | - Anthony E Samir
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | | | - Tina Kapur
- Department of Radiology, Brigham and Women's Hospital, Harvard Medical School, Boston, MA, United States
| | - Andrew J Goldsmith
- Department of Emergency Medicine, Lahey Hospital, University of Massachusetts Medical School, Burlington, MA, United States
| |
Collapse
|
17
|
Abdalla AR, Hageen AW, Saleh HH, Al-Azzawi O, Ghalab M, Harraz A, Eldoqsh BS, Elawady FE, Alhammadi AH, Elmorsy HH, Jano M, Elmasry M, Bahbah EI, Elgebaly A. Deep Learning Algorithms for the Detection of Suspicious Pigmented Skin Lesions in Primary Care Settings: A Systematic Review and Meta-Analysis. Cureus 2024; 16:e65122. [PMID: 39171046 PMCID: PMC11338545 DOI: 10.7759/cureus.65122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/22/2024] [Indexed: 08/23/2024] Open
Abstract
Early detection of suspicious pigmented skin lesions is crucial for improving the outcomes and survival rates of skin cancers. However, the accuracy of clinical diagnosis by primary care physicians (PCPs) is suboptimal, leading to unnecessary referrals and biopsies. In recent years, deep learning (DL) algorithms have shown promising results in the automated detection and classification of skin lesions. This systematic review and meta-analysis aimed to evaluate the diagnostic performance of DL algorithms for the detection of suspicious pigmented skin lesions in primary care settings. A comprehensive literature search was conducted using electronic databases, including PubMed, Scopus, IEEE Xplore, Cochrane Central Register of Controlled Trials (CENTRAL), and Web of Science. Data from eligible studies were extracted, including study characteristics, sample size, algorithm type, sensitivity, specificity, diagnostic odds ratio (DOR), positive likelihood ratio (PLR), negative likelihood ratio (NLR), and receiver operating characteristic curve analysis. Three studies were included. The results showed that DL algorithms had a high sensitivity (90%, 95% CI: 90-91%) and specificity (85%, 95% CI: 84-86%) for detecting suspicious pigmented skin lesions in primary care settings. Significant heterogeneity was observed in both sensitivity (p = 0.0062, I2 = 80.3%) and specificity (p < 0.001, I2 = 98.8%). The analysis of DOR and PLR further demonstrated the strong diagnostic performance of DL algorithms. The DOR was 26.39, indicating a strong overall diagnostic performance of DL algorithms. The PLR was 4.30, highlighting the ability of these algorithms to influence diagnostic outcomes positively. The NLR was 0.16, indicating that a negative test result decreased the odds of misdiagnosis. The area under the curve of DL algorithms was 0.95, indicating excellent discriminative ability in distinguishing between benign and malignant pigmented skin lesions. DL algorithms have the potential to significantly improve the detection of suspicious pigmented skin lesions in primary care settings. Our analysis showed that DL exhibited promising performance in the early detection of suspicious pigmented skin lesions. However, further studies are needed.
Collapse
Affiliation(s)
- Ahmed R Abdalla
- Vascular Surgery, Faculty of Medicine, Mansoura University, Mansoura, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Ahmed W Hageen
- Faculty of Medicine, Tanta University, Tanta, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Haneen H Saleh
- Faculty of Medicine, University of Jordan, Amman, JOR
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Omar Al-Azzawi
- Faculty of Pharmacy, İstinye University, İstanbul, TUR
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Mahmoud Ghalab
- Radiology Department, Kafrelsheikh University, Kafr El Sheikh, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Amani Harraz
- Faculty of Medicine, Alexandria University, Alexandria, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Bola S Eldoqsh
- Faculty of Medicine, Minia University, Minia, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Fatma E Elawady
- Department of Ophthalmology, Port Said Specialized Hospital of Ophthalmology, Port Said, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Ayman H Alhammadi
- Department of Radiology, Faculty of Medicine, Alexandria University, Alexandria, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Hesham Hassan Elmorsy
- Faculty of Pharmacy, Helwan University, Helwan, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Majd Jano
- Research Department, Syrian Society for Physicians and Pharmacists, Frankfurt, DEU
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Mohamed Elmasry
- Faculty of Medicine, Alexandria University, Alexandria, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Eshak I Bahbah
- Faculty of Medicine, Al-Azhar University, Damietta, EGY
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| | - Ahmed Elgebaly
- Smart Health Centre, University of East London, London, GBR
- Artificial Intelligence Research Group, MedDots Academy, Cairo, EGY
| |
Collapse
|
18
|
Collenne J, Monnier J, Iguernaissi R, Nawaf M, Richard MA, Grob JJ, Gaudy-Marqueste C, Dubuisson S, Merad D. Fusion between an Algorithm Based on the Characterization of Melanocytic Lesions' Asymmetry with an Ensemble of Convolutional Neural Networks for Melanoma Detection. J Invest Dermatol 2024; 144:1600-1607.e2. [PMID: 38296020 DOI: 10.1016/j.jid.2023.09.289] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2023] [Revised: 09/21/2023] [Accepted: 09/26/2023] [Indexed: 02/26/2024]
Abstract
Melanoma is still a major health problem worldwide. Early diagnosis is the first step toward reducing its mortality, but it remains a challenge even for experienced dermatologists. Although computer-aided systems have been developed to help diagnosis, the lack of insight into their predictions is still a significant limitation toward acceptance by the medical community. To tackle this issue, we designed handcrafted expert features representing color asymmetry within the lesions, which are parts of the approach used by dermatologists in their daily practice. These features are given to an artificial neural network classifying between nevi and melanoma. We compare our results with an ensemble of 7 state-of-the-art convolutional neural networks and merge the 2 approaches by computing the average prediction. Our experiments are done on a subset of the International Skin Imaging Collaboration 2019 dataset (6296 nevi, 1361 melanomas). The artificial neural network based on asymmetry achieved an area under the curve of 0.873, sensitivity of 90%, and specificity of 67%; the convolutional neural network approach achieved an area under the curve of 0.938, sensitivity of 91%, and specificity of 82%; and the fusion of both approaches achieved an area under the curve of 0.942, sensitivity of 92%, and specificity of 82%. Merging the knowledge of dermatologists with convolutional neural networks showed high performance for melanoma detection, encouraging collaboration between computer science and medical fields.
Collapse
Affiliation(s)
- Jules Collenne
- Computer Science and Systems Laboratory, CNRS UMR 7020, Aix-Marseille University, Marseille, France.
| | - Jilliana Monnier
- Computer Science and Systems Laboratory, CNRS UMR 7020, Aix-Marseille University, Marseille, France; Cancer Research Center of Marseille, Inserm UMR1068, CNRS UMR7258, Aix-Marseille University, Marseille, France; Dermatology and Skin Cancer Department, La Timone Hospital, Assistance Publique Hôpitaux de Marseille, Aix-Marseille University, Marseille, France
| | - Rabah Iguernaissi
- Computer Science and Systems Laboratory, CNRS UMR 7020, Aix-Marseille University, Marseille, France
| | - Motasem Nawaf
- Computer Science and Systems Laboratory, CNRS UMR 7020, Aix-Marseille University, Marseille, France
| | - Marie-Aleth Richard
- Dermatology and Skin Cancer Department, La Timone Hospital, Assistance Publique Hôpitaux de Marseille, Aix-Marseille University, Marseille, France
| | - Jean-Jacques Grob
- Cancer Research Center of Marseille, Inserm UMR1068, CNRS UMR7258, Aix-Marseille University, Marseille, France; Dermatology and Skin Cancer Department, La Timone Hospital, Assistance Publique Hôpitaux de Marseille, Aix-Marseille University, Marseille, France
| | - Caroline Gaudy-Marqueste
- Cancer Research Center of Marseille, Inserm UMR1068, CNRS UMR7258, Aix-Marseille University, Marseille, France; Dermatology and Skin Cancer Department, La Timone Hospital, Assistance Publique Hôpitaux de Marseille, Aix-Marseille University, Marseille, France
| | - Séverine Dubuisson
- Computer Science and Systems Laboratory, CNRS UMR 7020, Aix-Marseille University, Marseille, France
| | - Djamal Merad
- Computer Science and Systems Laboratory, CNRS UMR 7020, Aix-Marseille University, Marseille, France
| |
Collapse
|
19
|
Pundhir A, Sagar S, Singh P, Raman B. Echoes of images: multi-loss network for image retrieval in vision transformers. Med Biol Eng Comput 2024; 62:2037-2058. [PMID: 38436836 DOI: 10.1007/s11517-024-03055-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2023] [Accepted: 02/16/2024] [Indexed: 03/05/2024]
Abstract
This paper introduces a novel approach to enhance content-based image retrieval, validated on two benchmark datasets: ISIC-2017 and ISIC-2018. These datasets comprise skin lesion images that are crucial for innovations in skin cancer diagnosis and treatment. We advocate the use of pre-trained Vision Transformer (ViT), a relatively uncharted concept in the realm of image retrieval, particularly in medical scenarios. In contrast to the traditionally employed Convolutional Neural Networks (CNNs), our findings suggest that ViT offers a more comprehensive understanding of the image context, essential in medical imaging. We further incorporate a weighted multi-loss function, delving into various losses such as triplet loss, distillation loss, contrastive loss, and cross-entropy loss. Our exploration investigates the most resilient combination of these losses to create a robust multi-loss function, thus enhancing the robustness of the learned feature space and ameliorating the precision and recall in the retrieval process. Instead of using all the loss functions, the proposed multi-loss function utilizes the combination of only cross-entropy loss, triplet loss, and distillation loss and gains improvement of 6.52% and 3.45% for mean average precision over ISIC-2017 and ISIC-2018. Another innovation in our methodology is a two-branch network strategy, which concurrently boosts image retrieval and classification. Through our experiments, we underscore the effectiveness and the pitfalls of diverse loss configurations in image retrieval. Furthermore, our approach underlines the advantages of retrieval-based classification through majority voting rather than relying solely on the classification head, leading to enhanced prediction for melanoma - the most lethal type of skin cancer. Our results surpass existing state-of-the-art techniques on the ISIC-2017 and ISIC-2018 datasets by improving mean average precision by 1.01% and 4.36% respectively, emphasizing the efficacy and promise of Vision Transformers paired with our tailor-made weighted loss function, especially in medical contexts. The proposed approach's effectiveness is substantiated through thorough ablation studies and an array of quantitative and qualitative outcomes. To promote reproducibility and support forthcoming research, our source code will be accessible on GitHub.
Collapse
Affiliation(s)
- Anshul Pundhir
- Department of Computer Science and Engineering, Indian Institute of Technology, Roorkee, 247667, Uttarakhand, India.
| | - Shivam Sagar
- Department of Electrical Engineering, Indian Institute of Technology, Roorkee, 247667, Uttarakhand, India
| | - Pradeep Singh
- Department of Computer Science and Engineering, Indian Institute of Technology, Roorkee, 247667, Uttarakhand, India
| | - Balasubramanian Raman
- Department of Computer Science and Engineering, Indian Institute of Technology, Roorkee, 247667, Uttarakhand, India
| |
Collapse
|
20
|
Shahamatdar S, Saeed-Vafa D, Linsley D, Khalil F, Lovinger K, Li L, McLeod HT, Ramachandran S, Serre T. Deceptive learning in histopathology. Histopathology 2024; 85:116-132. [PMID: 38556922 PMCID: PMC11162337 DOI: 10.1111/his.15180] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Revised: 03/08/2024] [Accepted: 03/10/2024] [Indexed: 04/02/2024]
Abstract
AIMS Deep learning holds immense potential for histopathology, automating tasks that are simple for expert pathologists and revealing novel biology for tasks that were previously considered difficult or impossible to solve by eye alone. However, the extent to which the visual strategies learned by deep learning models in histopathological analysis are trustworthy or not has yet to be systematically analysed. Here, we systematically evaluate deep neural networks (DNNs) trained for histopathological analysis in order to understand if their learned strategies are trustworthy or deceptive. METHODS AND RESULTS We trained a variety of DNNs on a novel data set of 221 whole-slide images (WSIs) from lung adenocarcinoma patients, and evaluated their effectiveness at (1) molecular profiling of KRAS versus EGFR mutations, (2) determining the primary tissue of a tumour and (3) tumour detection. While DNNs achieved above-chance performance on molecular profiling, they did so by exploiting correlations between histological subtypes and mutations, and failed to generalise to a challenging test set obtained through laser capture microdissection (LCM). In contrast, DNNs learned robust and trustworthy strategies for determining the primary tissue of a tumour as well as detecting and localising tumours in tissue. CONCLUSIONS Our work demonstrates that DNNs hold immense promise for aiding pathologists in analysing tissue. However, they are also capable of achieving seemingly strong performance by learning deceptive strategies that leverage spurious correlations, and are ultimately unsuitable for research or clinical work. The framework we propose for model evaluation and interpretation is an important step towards developing reliable automated systems for histopathological analysis.
Collapse
Affiliation(s)
- Sahar Shahamatdar
- Center for Computational Molecular Biology, Brown University, Providence, RI, USA
- The Warren Alpert Medical School, Brown University, Providence, RI, USA
| | - Daryoush Saeed-Vafa
- Department of Anatomic Pathology, H. Lee Moffitt Cancer and Research Institute, Tampa, FL, USA
| | - Drew Linsley
- Carney Institute for Brain Science, Brown University, Providence, RI, USA
- Department of Cognitive Linguistic & Psychological Sciences, Brown University, Providence, RI, USA
| | - Farah Khalil
- Department of Anatomic Pathology, H. Lee Moffitt Cancer and Research Institute, Tampa, FL, USA
| | - Katherine Lovinger
- Department of Molecular Biology, H. Lee Moffitt Cancer and Research Institute, Tampa, FL, USA
| | - Lester Li
- University of Rochester, Rochester, NY, USA
| | | | - Sohini Ramachandran
- Center for Computational Molecular Biology, Brown University, Providence, RI, USA
- Department of Ecology, Evolution and Organismal Biology, Brown University, Providence, RI, USA
- The Data Science Initiative, Brown University, Providence, RI, USA
| | - Thomas Serre
- Carney Institute for Brain Science, Brown University, Providence, RI, USA
- Department of Cognitive Linguistic & Psychological Sciences, Brown University, Providence, RI, USA
| |
Collapse
|
21
|
Wang H, Jin Q, Li S, Liu S, Wang M, Song Z. A comprehensive survey on deep active learning in medical image analysis. Med Image Anal 2024; 95:103201. [PMID: 38776841 DOI: 10.1016/j.media.2024.103201] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Revised: 04/25/2024] [Accepted: 05/06/2024] [Indexed: 05/25/2024]
Abstract
Deep learning has achieved widespread success in medical image analysis, leading to an increasing demand for large-scale expert-annotated medical image datasets. Yet, the high cost of annotating medical images severely hampers the development of deep learning in this field. To reduce annotation costs, active learning aims to select the most informative samples for annotation and train high-performance models with as few labeled samples as possible. In this survey, we review the core methods of active learning, including the evaluation of informativeness and sampling strategy. For the first time, we provide a detailed summary of the integration of active learning with other label-efficient techniques, such as semi-supervised, self-supervised learning, and so on. We also summarize active learning works that are specifically tailored to medical image analysis. Additionally, we conduct a thorough comparative analysis of the performance of different AL methods in medical image analysis with experiments. In the end, we offer our perspectives on the future trends and challenges of active learning and its applications in medical image analysis. An accompanying paper list and code for the comparative analysis is available in https://github.com/LightersWang/Awesome-Active-Learning-for-Medical-Image-Analysis.
Collapse
Affiliation(s)
- Haoran Wang
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China
| | - Qiuye Jin
- Computational Bioscience Research Center (CBRC), King Abdullah University of Science and Technology (KAUST), Thuwal 23955, Saudi Arabia
| | - Shiman Li
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China
| | - Siyu Liu
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China
| | - Manning Wang
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China.
| | - Zhijian Song
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China.
| |
Collapse
|
22
|
Zhao J, Lui H, Kalia S, Lee TK, Zeng H. Improving skin cancer detection by Raman spectroscopy using convolutional neural networks and data augmentation. Front Oncol 2024; 14:1320220. [PMID: 38962264 PMCID: PMC11219827 DOI: 10.3389/fonc.2024.1320220] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2023] [Accepted: 05/23/2024] [Indexed: 07/05/2024] Open
Abstract
Background Our previous studies have demonstrated that Raman spectroscopy could be used for skin cancer detection with good sensitivity and specificity. The objective of this study is to determine if skin cancer detection can be further improved by combining deep neural networks and Raman spectroscopy. Patients and methods Raman spectra of 731 skin lesions were included in this study, containing 340 cancerous and precancerous lesions (melanoma, basal cell carcinoma, squamous cell carcinoma and actinic keratosis) and 391 benign lesions (melanocytic nevus and seborrheic keratosis). One-dimensional convolutional neural networks (1D-CNN) were developed for Raman spectral classification. The stratified samples were divided randomly into training (70%), validation (10%) and test set (20%), and were repeated 56 times using parallel computing. Different data augmentation strategies were implemented for the training dataset, including added random noise, spectral shift, spectral combination and artificially synthesized Raman spectra using one-dimensional generative adversarial networks (1D-GAN). The area under the receiver operating characteristic curve (ROC AUC) was used as a measure of the diagnostic performance. Conventional machine learning approaches, including partial least squares for discriminant analysis (PLS-DA), principal component and linear discriminant analysis (PC-LDA), support vector machine (SVM), and logistic regression (LR) were evaluated for comparison with the same data splitting scheme as the 1D-CNN. Results The ROC AUC of the test dataset based on the original training spectra were 0.886±0.022 (1D-CNN), 0.870±0.028 (PLS-DA), 0.875±0.033 (PC-LDA), 0.864±0.027 (SVM), and 0.525±0.045 (LR), which were improved to 0.909±0.021 (1D-CNN), 0.899±0.022 (PLS-DA), 0.895±0.022 (PC-LDA), 0.901±0.020 (SVM), and 0.897±0.021 (LR) respectively after augmentation of the training dataset (p<0.0001, Wilcoxon test). Paired analyses of 1D-CNN with conventional machine learning approaches showed that 1D-CNN had a 1-3% improvement (p<0.001, Wilcoxon test). Conclusions Data augmentation not only improved the performance of both deep neural networks and conventional machine learning techniques by 2-4%, but also improved the performance of the models on spectra with higher noise or spectral shifting. Convolutional neural networks slightly outperformed conventional machine learning approaches for skin cancer detection by Raman spectroscopy.
Collapse
Affiliation(s)
- Jianhua Zhao
- Photomedicine Institute, Department of Dermatology and Skin Science, University of British Columbia and Vancouver Coastal Health Research Institute, Vancouver, BC, Canada
- BC Cancer Research Institute, University of British Columbia, Vancouver, BC, Canada
| | - Harvey Lui
- Photomedicine Institute, Department of Dermatology and Skin Science, University of British Columbia and Vancouver Coastal Health Research Institute, Vancouver, BC, Canada
- BC Cancer Research Institute, University of British Columbia, Vancouver, BC, Canada
| | - Sunil Kalia
- Photomedicine Institute, Department of Dermatology and Skin Science, University of British Columbia and Vancouver Coastal Health Research Institute, Vancouver, BC, Canada
- BC Children’s Hospital Research Institute, Vancouver, BC, Canada
- Centre for Clinical Epidemiology and Evaluation, Vancouver Coastal Health Research Institute, Vancouver, BC, Canada
| | - Tim K. Lee
- Photomedicine Institute, Department of Dermatology and Skin Science, University of British Columbia and Vancouver Coastal Health Research Institute, Vancouver, BC, Canada
- BC Cancer Research Institute, University of British Columbia, Vancouver, BC, Canada
| | - Haishan Zeng
- Photomedicine Institute, Department of Dermatology and Skin Science, University of British Columbia and Vancouver Coastal Health Research Institute, Vancouver, BC, Canada
- BC Cancer Research Institute, University of British Columbia, Vancouver, BC, Canada
| |
Collapse
|
23
|
Escalé-Besa A, Vidal-Alaball J, Miró Catalina Q, Gracia VHG, Marin-Gomez FX, Fuster-Casanovas A. The Use of Artificial Intelligence for Skin Disease Diagnosis in Primary Care Settings: A Systematic Review. Healthcare (Basel) 2024; 12:1192. [PMID: 38921305 PMCID: PMC11202856 DOI: 10.3390/healthcare12121192] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2024] [Revised: 06/03/2024] [Accepted: 06/08/2024] [Indexed: 06/27/2024] Open
Abstract
The prevalence of dermatological conditions in primary care, coupled with challenges such as dermatologist shortages and rising consultation costs, highlights the need for innovative solutions. Artificial intelligence (AI) holds promise for improving the diagnostic analysis of skin lesion images, potentially enhancing patient care in primary settings. This systematic review following PRISMA guidelines examined primary studies (2012-2022) assessing AI algorithms' diagnostic accuracy for skin diseases in primary care. Studies were screened for eligibility based on their availability in the English language and exclusion criteria, with risk of bias evaluated using QUADAS-2. PubMed, Scopus, and Web of Science were searched. Fifteen studies (2019-2022), primarily from Europe and the USA, focusing on diagnostic accuracy were included. Sensitivity ranged from 58% to 96.1%, with accuracies varying from 0.41 to 0.93. AI applications encompassed triage and diagnostic support across diverse skin conditions in primary care settings, involving both patients and primary care professionals. While AI demonstrates potential for enhancing the accuracy of skin disease diagnostics in primary care, further research is imperative to address study heterogeneity and ensure algorithm reliability across diverse populations. Future investigations should prioritise robust dataset development and consider representative patient samples. Overall, AI may improve dermatological diagnosis in primary care, but careful consideration of algorithm limitations and implementation strategies is required.
Collapse
Affiliation(s)
- Anna Escalé-Besa
- Centre d’Atenció Primària Navàs-Balsareny, Institut Català de la Salut, 08670 Navàs, Spain;
- Health Promotion in Rural Areas Research Group, Gerència d’Atenció Primària i a la Comunitat de la Catalunya Central, Institut Català de la Salut, 08242 Manresa, Spain; (Q.M.C.); (F.X.M.-G.)
- Faculty of Medicine, University of Vic-Central University of Catalonia, 08500 Vic, Spain
| | - Josep Vidal-Alaball
- Health Promotion in Rural Areas Research Group, Gerència d’Atenció Primària i a la Comunitat de la Catalunya Central, Institut Català de la Salut, 08242 Manresa, Spain; (Q.M.C.); (F.X.M.-G.)
- Faculty of Medicine, University of Vic-Central University of Catalonia, 08500 Vic, Spain
- Unitat de Suport a la Recerca de la Catalunya Central, Fundació Institut Universitari per a la Recerca a l’Atenció Primària de Salut Jordi Gol i Gurina, 082424 Manresa, Spain;
| | - Queralt Miró Catalina
- Health Promotion in Rural Areas Research Group, Gerència d’Atenció Primària i a la Comunitat de la Catalunya Central, Institut Català de la Salut, 08242 Manresa, Spain; (Q.M.C.); (F.X.M.-G.)
- Unitat de Suport a la Recerca de la Catalunya Central, Fundació Institut Universitari per a la Recerca a l’Atenció Primària de Salut Jordi Gol i Gurina, 082424 Manresa, Spain;
| | | | - Francesc X. Marin-Gomez
- Health Promotion in Rural Areas Research Group, Gerència d’Atenció Primària i a la Comunitat de la Catalunya Central, Institut Català de la Salut, 08242 Manresa, Spain; (Q.M.C.); (F.X.M.-G.)
- Servei d’Atenció Primària Osona, Gerència Territorial de la Catalunya Central, Institut Català de La Salut, 08500 Vic, Spain
| | - Aïna Fuster-Casanovas
- Unitat de Suport a la Recerca de la Catalunya Central, Fundació Institut Universitari per a la Recerca a l’Atenció Primària de Salut Jordi Gol i Gurina, 082424 Manresa, Spain;
- eHealth Lab Research Group, School of Health Sciences and eHealth Centre, Universitat Oberta de Catalunya (UOC), 08018 Barcelona, Spain
| |
Collapse
|
24
|
Bryant AK, Zamora‐Resendiz R, Dai X, Morrow D, Lin Y, Jungles KM, Rae JM, Tate A, Pearson AN, Jiang R, Fritsche L, Lawrence TS, Zou W, Schipper M, Ramnath N, Yoo S, Crivelli S, Green MD. Artificial intelligence to unlock real-world evidence in clinical oncology: A primer on recent advances. Cancer Med 2024; 13:e7253. [PMID: 38899720 PMCID: PMC11187737 DOI: 10.1002/cam4.7253] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2023] [Revised: 02/05/2024] [Accepted: 04/28/2024] [Indexed: 06/21/2024] Open
Abstract
PURPOSE Real world evidence is crucial to understanding the diffusion of new oncologic therapies, monitoring cancer outcomes, and detecting unexpected toxicities. In practice, real world evidence is challenging to collect rapidly and comprehensively, often requiring expensive and time-consuming manual case-finding and annotation of clinical text. In this Review, we summarise recent developments in the use of artificial intelligence to collect and analyze real world evidence in oncology. METHODS We performed a narrative review of the major current trends and recent literature in artificial intelligence applications in oncology. RESULTS Artificial intelligence (AI) approaches are increasingly used to efficiently phenotype patients and tumors at large scale. These tools also may provide novel biological insights and improve risk prediction through multimodal integration of radiographic, pathological, and genomic datasets. Custom language processing pipelines and large language models hold great promise for clinical prediction and phenotyping. CONCLUSIONS Despite rapid advances, continued progress in computation, generalizability, interpretability, and reliability as well as prospective validation are needed to integrate AI approaches into routine clinical care and real-time monitoring of novel therapies.
Collapse
Affiliation(s)
- Alex K. Bryant
- Department of Radiation OncologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
- Department of Radiation Oncology, Veterans Affairs Ann Arbor Healthcare SystemAnn ArborMichiganUSA
| | - Rafael Zamora‐Resendiz
- Applied Mathematics and Computational Research Division, Lawrence Berkeley National LaboratoryBerkeleyCaliforniaUSA
| | - Xin Dai
- Computational Science Initiative, Brookhaven National LaboratoryUptonNew YorkUSA
| | - Destinee Morrow
- Applied Mathematics and Computational Research Division, Lawrence Berkeley National LaboratoryBerkeleyCaliforniaUSA
| | - Yuewei Lin
- Computational Science Initiative, Brookhaven National LaboratoryUptonNew YorkUSA
| | - Kassidy M. Jungles
- Department of PharmacologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
| | - James M. Rae
- Department of PharmacologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
- Department of Internal MedicineUniversity of Michigan School of MedicineAnn ArborMichiganUSA
| | - Akshay Tate
- Department of Radiation OncologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
| | - Ashley N. Pearson
- Department of Radiation OncologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
| | - Ralph Jiang
- Department of Radiation OncologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
- Department of StatisticsUniversity of MichiganAnn ArborMichiganUSA
| | - Lars Fritsche
- Department of StatisticsUniversity of MichiganAnn ArborMichiganUSA
| | - Theodore S. Lawrence
- Department of Radiation OncologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
| | - Weiping Zou
- Department of StatisticsUniversity of MichiganAnn ArborMichiganUSA
- Center of Excellence for Cancer Immunology and ImmunotherapyUniversity of Michigan Rogel Cancer CenterAnn ArborMichiganUSA
- Department of PathologyUniversity of MichiganAnn ArborMichiganUSA
- Graduate Program in ImmunologyUniversity of MichiganAnn ArborMichiganUSA
| | - Matthew Schipper
- Department of Radiation OncologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
- Department of PharmacologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
| | - Nithya Ramnath
- Division of Hematology Oncology, Department of MedicineUniversity of Michigan School of MedicineAnn ArborMichiganUSA
- Division of Hematology Oncology, Department of MedicineVeterans Affairs Ann Arbor Healthcare SystemAnn ArborMichiganUSA
| | - Shinjae Yoo
- Computational Science Initiative, Brookhaven National LaboratoryUptonNew YorkUSA
| | - Silvia Crivelli
- Applied Mathematics and Computational Research Division, Lawrence Berkeley National LaboratoryBerkeleyCaliforniaUSA
| | - Michael D. Green
- Department of Radiation OncologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
- Department of Radiation Oncology, Veterans Affairs Ann Arbor Healthcare SystemAnn ArborMichiganUSA
- Graduate Program in ImmunologyUniversity of MichiganAnn ArborMichiganUSA
- Graduate Program in Cancer BiologyUniversity of MichiganAnn ArborMichiganUSA
- Department of Microbiology and ImmunologyUniversity of Michigan School of MedicineAnn ArborMichiganUSA
| |
Collapse
|
25
|
Primiero CA, Rezze GG, Caffery LJ, Carrera C, Podlipnik S, Espinosa N, Puig S, Janda M, Soyer HP, Malvehy J. A Narrative Review: Opportunities and Challenges in Artificial Intelligence Skin Image Analyses Using Total Body Photography. J Invest Dermatol 2024; 144:1200-1207. [PMID: 38231164 DOI: 10.1016/j.jid.2023.11.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Revised: 09/19/2023] [Accepted: 11/09/2023] [Indexed: 01/18/2024]
Abstract
Artificial intelligence (AI) algorithms for skin lesion classification have reported accuracy at par with and even outperformance of expert dermatologists in experimental settings. However, the majority of algorithms do not represent real-world clinical approach where skin phenotype and clinical background information are considered. We review the current state of AI for skin lesion classification and present opportunities and challenges when applied to total body photography (TBP). AI in TBP analysis presents opportunities for intrapatient assessment of skin phenotype and holistic risk assessment by incorporating patient-level metadata, although challenges exist for protecting patient privacy in algorithm development and improving explainable AI methods.
Collapse
Affiliation(s)
- Clare A Primiero
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica - Institut d'Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain; Dermatology Research Centre, Frazer Institute, The University of Queensland, Brisbane, Australia
| | - Gisele Gargantini Rezze
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica - Institut d'Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain
| | - Liam J Caffery
- Dermatology Research Centre, Frazer Institute, The University of Queensland, Brisbane, Australia; Centre of Health Services Research, Faculty of Medicine, The University of Queensland, Brisbane, Australia; Centre for Online Health, The University of Queensland, Brisbane, Australia
| | - Cristina Carrera
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica - Institut d'Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain; Medicine Department, University of Barcelona, Barcelona, Spain; CIBER de Enfermedades raras, Instituto de Salud Carlos III, Barcelona, Spain
| | - Sebastian Podlipnik
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica - Institut d'Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain; CIBER de Enfermedades raras, Instituto de Salud Carlos III, Barcelona, Spain
| | - Natalia Espinosa
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica - Institut d'Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain
| | - Susana Puig
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica - Institut d'Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain; Medicine Department, University of Barcelona, Barcelona, Spain; CIBER de Enfermedades raras, Instituto de Salud Carlos III, Barcelona, Spain
| | - Monika Janda
- Centre of Health Services Research, Faculty of Medicine, The University of Queensland, Brisbane, Australia
| | - H Peter Soyer
- Dermatology Research Centre, Frazer Institute, The University of Queensland, Brisbane, Australia; Dermatology Department, Princess Alexandra Hospital, Brisbane, Australia
| | - Josep Malvehy
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica - Institut d'Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain; Medicine Department, University of Barcelona, Barcelona, Spain; CIBER de Enfermedades raras, Instituto de Salud Carlos III, Barcelona, Spain.
| |
Collapse
|
26
|
Scott IA, van der Vegt A, Lane P, McPhail S, Magrabi F. Achieving large-scale clinician adoption of AI-enabled decision support. BMJ Health Care Inform 2024; 31:e100971. [PMID: 38816209 PMCID: PMC11141172 DOI: 10.1136/bmjhci-2023-100971] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2023] [Accepted: 05/15/2024] [Indexed: 06/01/2024] Open
Abstract
Computerised decision support (CDS) tools enabled by artificial intelligence (AI) seek to enhance accuracy and efficiency of clinician decision-making at the point of care. Statistical models developed using machine learning (ML) underpin most current tools. However, despite thousands of models and hundreds of regulator-approved tools internationally, large-scale uptake into routine clinical practice has proved elusive. While underdeveloped system readiness and investment in AI/ML within Australia and perhaps other countries are impediments, clinician ambivalence towards adopting these tools at scale could be a major inhibitor. We propose a set of principles and several strategic enablers for obtaining broad clinician acceptance of AI/ML-enabled CDS tools.
Collapse
Affiliation(s)
- Ian A Scott
- Internal Medicine and Clinical Epidemiology, Princess Alexandra Hospital, Brisbane, Queensland, Australia
- Centre for Health Services Research, The University of Queensland Faculty of Medicine and Biomedical Sciences, Brisbane, Queensland, Australia
| | - Anton van der Vegt
- Digital Health Centre, The University of Queensland Faculty of Medicine and Biomedical Sciences, Herston, Queensland, Australia
| | - Paul Lane
- Safety, Quality and Innovation, The Prince Charles Hospital, Brisbane, Queensland, Australia
| | - Steven McPhail
- Australian Centre for Health Services Innovation, Queensland University of Technology Faculty of Health, Brisbane, Queensland, Australia
| | - Farah Magrabi
- Macquarie University, Sydney, New South Wales, Australia
| |
Collapse
|
27
|
Hasan E, Duhaime E, Trueblood JS. Boosting wisdom of the crowd for medical image annotation using training performance and task features. Cogn Res Princ Implic 2024; 9:31. [PMID: 38763994 PMCID: PMC11102897 DOI: 10.1186/s41235-024-00558-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2023] [Accepted: 04/29/2024] [Indexed: 05/21/2024] Open
Abstract
A crucial bottleneck in medical artificial intelligence (AI) is high-quality labeled medical datasets. In this paper, we test a large variety of wisdom of the crowd algorithms to label medical images that were initially classified by individuals recruited through an app-based platform. Individuals classified skin lesions from the International Skin Lesion Challenge 2018 into 7 different categories. There was a large dispersion in the geographical location, experience, training, and performance of the recruited individuals. We tested several wisdom of the crowd algorithms of varying complexity from a simple unweighted average to more complex Bayesian models that account for individual patterns of errors. Using a switchboard analysis, we observe that the best-performing algorithms rely on selecting top performers, weighting decisions by training accuracy, and take into account the task environment. These algorithms far exceed expert performance. We conclude by discussing the implications of these approaches for the development of medical AI.
Collapse
Affiliation(s)
- Eeshan Hasan
- Department of Psychological and Brain Sciences, Indiana University, 1101 E. 10th St., Bloomington, IN, 47405-7007, USA.
- Cognitive Science Program, Indiana University, Bloomington, USA.
| | | | - Jennifer S Trueblood
- Department of Psychological and Brain Sciences, Indiana University, 1101 E. 10th St., Bloomington, IN, 47405-7007, USA.
- Cognitive Science Program, Indiana University, Bloomington, USA.
| |
Collapse
|
28
|
Salinas MP, Sepúlveda J, Hidalgo L, Peirano D, Morel M, Uribe P, Rotemberg V, Briones J, Mery D, Navarrete-Dechent C. A systematic review and meta-analysis of artificial intelligence versus clinicians for skin cancer diagnosis. NPJ Digit Med 2024; 7:125. [PMID: 38744955 PMCID: PMC11094047 DOI: 10.1038/s41746-024-01103-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2023] [Accepted: 04/04/2024] [Indexed: 05/16/2024] Open
Abstract
Scientific research of artificial intelligence (AI) in dermatology has increased exponentially. The objective of this study was to perform a systematic review and meta-analysis to evaluate the performance of AI algorithms for skin cancer classification in comparison to clinicians with different levels of expertise. Based on PRISMA guidelines, 3 electronic databases (PubMed, Embase, and Cochrane Library) were screened for relevant articles up to August 2022. The quality of the studies was assessed using QUADAS-2. A meta-analysis of sensitivity and specificity was performed for the accuracy of AI and clinicians. Fifty-three studies were included in the systematic review, and 19 met the inclusion criteria for the meta-analysis. Considering all studies and all subgroups of clinicians, we found a sensitivity (Sn) and specificity (Sp) of 87.0% and 77.1% for AI algorithms, respectively, and a Sn of 79.78% and Sp of 73.6% for all clinicians (overall); differences were statistically significant for both Sn and Sp. The difference between AI performance (Sn 92.5%, Sp 66.5%) vs. generalists (Sn 64.6%, Sp 72.8%), was greater, when compared with expert clinicians. Performance between AI algorithms (Sn 86.3%, Sp 78.4%) vs expert dermatologists (Sn 84.2%, Sp 74.4%) was clinically comparable. Limitations of AI algorithms in clinical practice should be considered, and future studies should focus on real-world settings, and towards AI-assistance.
Collapse
Affiliation(s)
- Maria Paz Salinas
- Department of Dermatology, Escuela de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Javiera Sepúlveda
- Department of Dermatology, Escuela de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Leonel Hidalgo
- Department of Dermatology, Escuela de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Dominga Peirano
- Department of Dermatology, Escuela de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Macarena Morel
- Universidad Catolica-Evidence Center, Cochrane Chile Associated Center, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Pablo Uribe
- Department of Dermatology, Escuela de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile
- Melanoma and Skin Cancer Unit, Escuela de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Veronica Rotemberg
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, NY, USA
| | - Juan Briones
- Department of Oncology, Escuela de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Domingo Mery
- Department of Computer Science, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Cristian Navarrete-Dechent
- Department of Dermatology, Escuela de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile.
- Melanoma and Skin Cancer Unit, Escuela de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile.
| |
Collapse
|
29
|
Giavina-Bianchi M, Ko J. Editorial: Artificial intelligence in cutaneous lesions: where do we stand and what is next? Front Med (Lausanne) 2024; 11:1420152. [PMID: 38779620 PMCID: PMC11109431 DOI: 10.3389/fmed.2024.1420152] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Accepted: 04/30/2024] [Indexed: 05/25/2024] Open
Affiliation(s)
- Mara Giavina-Bianchi
- Medical Image Research Department, Hospital Israelita Albert Einstein, São Paulo, Brazil
| | - Justin Ko
- Clinical Dermatology, School of Medicine, Stanford University, Stanford, CA, United States
| |
Collapse
|
30
|
Minagawa A, Koga H, Matsunaga K, Hamada A, Okuyama R. Difference in computer-assistance effects on human doctor accuracy in diagnosing acral and non-acral pigmented skin tumors. J Dermatol 2024; 51:735-737. [PMID: 38634738 DOI: 10.1111/1346-8138.17240] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2023] [Revised: 03/12/2024] [Accepted: 04/04/2024] [Indexed: 04/19/2024]
Affiliation(s)
- Akane Minagawa
- Department of Dermatology, Shinshu University School of Medicine, Nagano, Japan
| | - Hiroshi Koga
- Department of Dermatology, Shinshu University School of Medicine, Nagano, Japan
| | | | | | - Ryuhei Okuyama
- Department of Dermatology, Shinshu University School of Medicine, Nagano, Japan
| |
Collapse
|
31
|
Ingvar Å, Oloruntoba A, Sashindranath M, Miller R, Soyer HP, Guitera P, Caccetta T, Shumack S, Abbott L, Arnold C, Lawn C, Button-Sloan A, Janda M, Mar V. Minimum labelling requirements for dermatology artificial intelligence-based Software as Medical Device (SaMD): A consensus statement. Australas J Dermatol 2024; 65:e21-e29. [PMID: 38419186 DOI: 10.1111/ajd.14222] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/26/2023] [Accepted: 01/21/2024] [Indexed: 03/02/2024]
Abstract
BACKGROUND/OBJECTIVES Artificial intelligence (AI) holds remarkable potential to improve care delivery in dermatology. End users (health professionals and general public) of AI-based Software as Medical Devices (SaMD) require relevant labelling information to ensure that these devices can be used appropriately. Currently, there are no clear minimum labelling requirements for dermatology AI-based SaMDs. METHODS Common labelling recommendations for AI-based SaMD identified in a recent literature review were evaluated by an Australian expert panel in digital health and dermatology via a modified Delphi consensus process. A nine-point Likert scale was used to indicate importance of 10 items, and voting was conducted to determine the specific characteristics to include for some items. Consensus was achieved when more than 75% of the experts agreed that inclusion of information was necessary. RESULTS There was robust consensus supporting inclusion of all proposed items as minimum labelling requirements; indication for use, intended user, training and test data sets, algorithm design, image processing techniques, clinical validation, performance metrics, limitations, updates and adverse events. Nearly all suggested characteristics of the labelling items received endorsement, except for some characteristics related to performance metrics. Moreover, there was consensus that uniform labelling criteria should apply across all AI categories and risk classes set out by the Therapeutic Goods Administration. CONCLUSIONS This study provides critical evidence for setting labelling standards by the Therapeutic Goods Administration to safeguard patients, health professionals, consumers, industry, and regulatory bodies from AI-based dermatology SaMDs that do not currently provide adequate information about how they were developed and tested.
Collapse
Affiliation(s)
- Åsa Ingvar
- Victorian Melanoma Service, Alfred Health, Melbourne, Victoria, Australia
- School of Public Health and Preventive Medicine, Monash University, Melbourne, Victoria, Australia
- Department of Dermatology, Skåne University Hospital, Lund, Sweden
- Department of Clinical Sciences, Lund University, Lund, Sweden
| | | | - Maithili Sashindranath
- School of Public Health and Preventive Medicine, Monash University, Melbourne, Victoria, Australia
| | - Robert Miller
- Australasian College of Dermatologists, Sydney, Australia
| | - H Peter Soyer
- Australasian College of Dermatologists, Sydney, Australia
- Dermatology Research Centre, Frazer Institute, The University of Queensland, Brisbane, Queensland, Australia
| | - Pascale Guitera
- Australasian College of Dermatologists, Sydney, Australia
- Faculty of Medicine and Health, The University of Sydney, Sydney, New South Wales, Australia
- Sydney Melanoma Diagnostic Centre, Royal Prince Alfred Hospital, Camperdown, Victoria, Australia
- Melanoma Institute Australia, The University of Sydney, Sydney, New South Wales, Australia
| | - Tony Caccetta
- Australasian College of Dermatologists, Sydney, Australia
- Perth Dermatology Clinic, Perth, Western Australia, Australia
| | - Stephen Shumack
- Australasian College of Dermatologists, Sydney, Australia
- Royal North Shore Hospital of Sydney, Sydney, New South Wales, Australia
| | - Lisa Abbott
- Australasian College of Dermatologists, Sydney, Australia
- Faculty of Medicine and Health, The University of Sydney, Sydney, New South Wales, Australia
- The Skin Hospital, Sydney, New South Wales, Australia
| | - Chris Arnold
- BioGrid Australia Ltd, Melbourne, Australia
- Hodgson Associates, Melbourne, Australia
- Australasian Society of Cosmetic Dermatologists, Melbourne, Australia
| | - Craig Lawn
- Melanoma Institute Australia, The University of Sydney, Sydney, New South Wales, Australia
- Centre of Excellence in Melanoma Imaging, Brisbane, Queensland, Australia
| | | | - Monika Janda
- Australasian College of Dermatologists, Sydney, Australia
- Dermatology Research Centre, Frazer Institute, The University of Queensland, Brisbane, Queensland, Australia
- Centre for Health Services Research, The University of Queensland, Brisbane, Queensland, Australia
| | - Victoria Mar
- Victorian Melanoma Service, Alfred Health, Melbourne, Victoria, Australia
- School of Public Health and Preventive Medicine, Monash University, Melbourne, Victoria, Australia
- Australasian College of Dermatologists, Sydney, Australia
| |
Collapse
|
32
|
Santomartino SM, Putman K, Beheshtian E, Parekh VS, Yi PH. Evaluating the Robustness of a Deep Learning Bone Age Algorithm to Clinical Image Variation Using Computational Stress Testing. Radiol Artif Intell 2024; 6:e230240. [PMID: 38477660 PMCID: PMC11140516 DOI: 10.1148/ryai.230240] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Revised: 02/08/2024] [Accepted: 02/26/2024] [Indexed: 03/14/2024]
Abstract
Purpose To evaluate the robustness of an award-winning bone age deep learning (DL) model to extensive variations in image appearance. Materials and Methods In December 2021, the DL bone age model that won the 2017 RSNA Pediatric Bone Age Challenge was retrospectively evaluated using the RSNA validation set (1425 pediatric hand radiographs; internal test set in this study) and the Digital Hand Atlas (DHA) (1202 pediatric hand radiographs; external test set). Each test image underwent seven types of transformations (rotations, flips, brightness, contrast, inversion, laterality marker, and resolution) to represent a range of image appearances, many of which simulate real-world variations. Computational "stress tests" were performed by comparing the model's predictions on baseline and transformed images. Mean absolute differences (MADs) of predicted bone ages compared with radiologist-determined ground truth on baseline versus transformed images were compared using Wilcoxon signed rank tests. The proportion of clinically significant errors (CSEs) was compared using McNemar tests. Results There was no evidence of a difference in MAD of the model on the two baseline test sets (RSNA = 6.8 months, DHA = 6.9 months; P = .05), indicating good model generalization to external data. Except for the RSNA dataset images with an appended radiologic laterality marker (P = .86), there were significant differences in MAD for both the DHA and RSNA datasets among other transformation groups (rotations, flips, brightness, contrast, inversion, and resolution). There were significant differences in proportion of CSEs for 57% of the image transformations (19 of 33) performed on the DHA dataset. Conclusion Although an award-winning pediatric bone age DL model generalized well to curated external images, it had inconsistent predictions on images that had undergone simple transformations reflective of several real-world variations in image appearance. Keywords: Pediatrics, Hand, Convolutional Neural Network, Radiography Supplemental material is available for this article. © RSNA, 2024 See also commentary by Faghani and Erickson in this issue.
Collapse
Affiliation(s)
- Samantha M Santomartino
- From the Drexel University College of Medicine, Philadelphia, Pa (S.M.S.); University of Maryland Medical Intelligent Imaging (UM2ii) Center, Department of Diagnostic Radiology and Nuclear Medicine, University of Maryland School of Medicine, 670 W Baltimore St, 1st Fl, Room 1172, Baltimore, MD 21201 (S.M.S., K.P., E.B., V.S.P., P.H.Y.); and Malone Center for Engineering in Healthcare, Johns Hopkins University, Baltimore, Md (P.H.Y.)
| | - Kristin Putman
- From the Drexel University College of Medicine, Philadelphia, Pa (S.M.S.); University of Maryland Medical Intelligent Imaging (UM2ii) Center, Department of Diagnostic Radiology and Nuclear Medicine, University of Maryland School of Medicine, 670 W Baltimore St, 1st Fl, Room 1172, Baltimore, MD 21201 (S.M.S., K.P., E.B., V.S.P., P.H.Y.); and Malone Center for Engineering in Healthcare, Johns Hopkins University, Baltimore, Md (P.H.Y.)
| | - Elham Beheshtian
- From the Drexel University College of Medicine, Philadelphia, Pa (S.M.S.); University of Maryland Medical Intelligent Imaging (UM2ii) Center, Department of Diagnostic Radiology and Nuclear Medicine, University of Maryland School of Medicine, 670 W Baltimore St, 1st Fl, Room 1172, Baltimore, MD 21201 (S.M.S., K.P., E.B., V.S.P., P.H.Y.); and Malone Center for Engineering in Healthcare, Johns Hopkins University, Baltimore, Md (P.H.Y.)
| | - Vishwa S Parekh
- From the Drexel University College of Medicine, Philadelphia, Pa (S.M.S.); University of Maryland Medical Intelligent Imaging (UM2ii) Center, Department of Diagnostic Radiology and Nuclear Medicine, University of Maryland School of Medicine, 670 W Baltimore St, 1st Fl, Room 1172, Baltimore, MD 21201 (S.M.S., K.P., E.B., V.S.P., P.H.Y.); and Malone Center for Engineering in Healthcare, Johns Hopkins University, Baltimore, Md (P.H.Y.)
| | - Paul H Yi
- From the Drexel University College of Medicine, Philadelphia, Pa (S.M.S.); University of Maryland Medical Intelligent Imaging (UM2ii) Center, Department of Diagnostic Radiology and Nuclear Medicine, University of Maryland School of Medicine, 670 W Baltimore St, 1st Fl, Room 1172, Baltimore, MD 21201 (S.M.S., K.P., E.B., V.S.P., P.H.Y.); and Malone Center for Engineering in Healthcare, Johns Hopkins University, Baltimore, Md (P.H.Y.)
| |
Collapse
|
33
|
Winkler JK, Kommoss KS, Toberer F, Enk A, Maul LV, Navarini AA, Hudson J, Salerni G, Rosenberger A, Haenssle HA. Performance of an automated total body mapping algorithm to detect melanocytic lesions of clinical relevance. Eur J Cancer 2024; 202:114026. [PMID: 38547776 DOI: 10.1016/j.ejca.2024.114026] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2024] [Revised: 03/11/2024] [Accepted: 03/14/2024] [Indexed: 04/21/2024]
Abstract
IMPORTANCE Total body photography for skin cancer screening is a well-established tool allowing documentation and follow-up of the entire skin surface. Artificial intelligence-based systems are increasingly applied for automated lesion detection and diagnosis. DESIGN AND PATIENTS In this prospective observational international multicentre study experienced dermatologists performed skin cancer screenings and identified clinically relevant melanocytic lesions (CRML, requiring biopsy or observation). Additionally, patients received 2D automated total body mapping (ATBM) with automated lesion detection (ATBM master, Fotofinder Systems GmbH). Primary endpoint was the percentage of CRML detected by the bodyscan software. Secondary endpoints included the percentage of correctly identified "new" and "changed" lesions during follow-up examinations. RESULTS At baseline, dermatologists identified 1075 CRML in 236 patients and 999 CRML (92.9%) were also detected by the automated software. During follow-up examinations dermatologists identified 334 CRMLs in 55 patients, with 323 (96.7%) also being detected by ATBM with automated lesions detection. Moreover, all new (n = 13) or changed CRML (n = 24) during follow-up were detected by the software. Average time requirements per baseline examination was 14.1 min (95% CI [12.8-15.5]). Subgroup analysis of undetected lesions revealed either technical (e.g. covering by clothing, hair) or lesion-specific reasons (e.g. hypopigmentation, palmoplantar sites). CONCLUSIONS ATBM with lesion detection software correctly detected the vast majority of CRML and new or changed CRML during follow-up examinations in a favourable amount of time. Our prospective international study underlines that automated lesion detection in TBP images is feasible, which is of relevance for developing AI-based skin cancer screenings.
Collapse
Affiliation(s)
- Julia K Winkler
- Department of Dermatology, University of Heidelberg, Heidelberg, Germany.
| | | | - Ferdinand Toberer
- Department of Dermatology, University of Heidelberg, Heidelberg, Germany
| | - Alexander Enk
- Department of Dermatology, University of Heidelberg, Heidelberg, Germany
| | - Lara V Maul
- Department of Dermatology, University Hospital of Basel, Basel, Switzerland
| | | | - Jeremy Hudson
- North Queensland Skin Centre, Townsville, Queensland, Australia
| | - Gabriel Salerni
- Department of Dermatology, Hospital Provincial del Centenario de Rosario- Universidad Nacional de Rosario, Rosario, Argentina
| | - Albert Rosenberger
- Institute of Genetic Epidemiology, University Medical Center, Georg-August University of Goettingen, Goettingen, Germany
| | - Holger A Haenssle
- Department of Dermatology, University of Heidelberg, Heidelberg, Germany
| |
Collapse
|
34
|
Strzelecki M, Kociołek M, Strąkowska M, Kozłowski M, Grzybowski A, Szczypiński PM. Artificial intelligence in the detection of skin cancer: State of the art. Clin Dermatol 2024; 42:280-295. [PMID: 38181888 DOI: 10.1016/j.clindermatol.2023.12.022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/07/2024]
Abstract
The incidence of melanoma is increasing rapidly. This cancer has a good prognosis if detected early. For this reason, various systems of skin lesion image analysis, which support imaging diagnostics of this neoplasm, are developing very dynamically. To detect and recognize neoplastic lesions, such systems use various artificial intelligence (AI) algorithms. This area of computer science applications has recently undergone dynamic development, abounding in several solutions that are effective tools supporting diagnosticians in many medical specialties. In this contribution, a number of applications of different classes of AI algorithms for the detection of this skin melanoma are presented and evaluated. Both classic systems based on the analysis of dermatoscopic images as well as total body systems, enabling the analysis of the patient's whole body to detect moles and pathologic changes, are discussed. These increasingly popular applications that allow the analysis of lesion images using smartphones are also described. The quantitative evaluation of the discussed systems with particular emphasis on the method of validation of the implemented algorithms is presented. The advantages and limitations of AI in the analysis of lesion images are also discussed, and problems requiring a solution for more effective use of AI in dermatology are identified.
Collapse
Affiliation(s)
- Michał Strzelecki
- Institute of Electronics, Lodz University of Technology, Łódź, Poland.
| | - Marcin Kociołek
- Institute of Electronics, Lodz University of Technology, Łódź, Poland
| | - Maria Strąkowska
- Institute of Electronics, Lodz University of Technology, Łódź, Poland
| | - Michał Kozłowski
- Department of Mechatronics and Technical and IT Education, Faculty of Technical Science, University of Warmia and Mazury, Olsztyn, Poland
| | - Andrzej Grzybowski
- Institute for Research in Ophthalmology, Foundation for Ophthalmology Development, Poznan, Poland
| | | |
Collapse
|
35
|
Niraula D, Cuneo KC, Dinov ID, Gonzalez BD, Jamaluddin JB, Jin JJ, Luo Y, Matuszak MM, Ten Haken RK, Bryant AK, Dilling TJ, Dykstra MP, Frakes JM, Liveringhouse CL, Miller SR, Mills MN, Palm RF, Regan SN, Rishi A, Torres-Roca JF, Yu HHM, El Naqa I. Intricacies of Human-AI Interaction in Dynamic Decision-Making for Precision Oncology: A Case Study in Response-Adaptive Radiotherapy. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2024:2024.04.27.24306434. [PMID: 38746238 PMCID: PMC11092730 DOI: 10.1101/2024.04.27.24306434] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2024]
Abstract
Background Adaptive treatment strategies that can dynamically react to individual cancer progression can provide effective personalized care. Longitudinal multi-omics information, paired with an artificially intelligent clinical decision support system (AI-CDSS) can assist clinicians in determining optimal therapeutic options and treatment adaptations. However, AI-CDSS is not perfectly accurate, as such, clinicians' over/under reliance on AI may lead to unintended consequences, ultimately failing to develop optimal strategies. To investigate such collaborative decision-making process, we conducted a Human-AI interaction case study on response-adaptive radiotherapy (RT). Methods We designed and conducted a two-phase study for two disease sites and two treatment modalities-adaptive RT for non-small cell lung cancer (NSCLC) and adaptive stereotactic body RT for hepatocellular carcinoma (HCC)-in which clinicians were asked to consider mid-treatment modification of the dose per fraction for a number of retrospective cancer patients without AI-support (Unassisted Phase) and with AI-assistance (AI-assisted Phase). The AI-CDSS graphically presented trade-offs in tumor control and the likelihood of toxicity to organs at risk, provided an optimal recommendation, and associated model uncertainties. In addition, we asked for clinicians' decision confidence level and trust level in individual AI recommendations and encouraged them to provide written remarks. We enrolled 13 evaluators (radiation oncology physicians and residents) from two medical institutions located in two different states, out of which, 4 evaluators volunteered in both NSCLC and HCC studies, resulting in a total of 17 completed evaluations (9 NSCLC, and 8 HCC). To limit the evaluation time to under an hour, we selected 8 treated patients for NSCLC and 9 for HCC, resulting in a total of 144 sets of evaluations (72 from NSCLC and 72 from HCC). Evaluation for each patient consisted of 8 required inputs and 2 optional remarks, resulting in up to a total of 1440 data points. Results AI-assistance did not homogeneously influence all experts and clinical decisions. From NSCLC cohort, 41 (57%) decisions and from HCC cohort, 34 (47%) decisions were adjusted after AI assistance. Two evaluations (12%) from the NSCLC cohort had zero decision adjustments, while the remaining 15 (88%) evaluations resulted in at least two decision adjustments. Decision adjustment level positively correlated with dissimilarity in decision-making with AI [NSCLC: ρ = 0.53 ( p < 0.001); HCC: ρ = 0.60 ( p < 0.001)] indicating that evaluators adjusted their decision closer towards AI recommendation. Agreement with AI-recommendation positively correlated with AI Trust Level [NSCLC: ρ = 0.59 ( p < 0.001); HCC: ρ = 0.7 ( p < 0.001)] indicating that evaluators followed AI's recommendation if they agreed with that recommendation. The correlation between decision confidence changes and decision adjustment level showed an opposite trend [NSCLC: ρ = -0.24 ( p = 0.045), HCC: ρ = 0.28 ( p = 0.017)] reflecting the difference in behavior due to underlying differences in disease type and treatment modality. Decision confidence positively correlated with the closeness of decisions to the standard of care (NSCLC: 2 Gy/fx; HCC: 10 Gy/fx) indicating that evaluators were generally more confident in prescribing dose fractionations more similar to those used in standard clinical practice. Inter-evaluator agreement increased with AI-assistance indicating that AI-assistance can decrease inter-physician variability. The majority of decisions were adjusted to achieve higher tumor control in NSCLC and lower normal tissue complications in HCC. Analysis of evaluators' remarks indicated concerns for organs at risk and RT outcome estimates as important decision-making factors. Conclusions Human-AI interaction depends on the complex interrelationship between expert's prior knowledge and preferences, patient's state, disease site, treatment modality, model transparency, and AI's learned behavior and biases. The collaborative decision-making process can be summarized as follows: (i) some clinicians may not believe in an AI system, completely disregarding its recommendation, (ii) some clinicians may believe in the AI system but will critically analyze its recommendations on a case-by-case basis; (iii) when a clinician finds that the AI recommendation indicates the possibility for better outcomes they will adjust their decisions accordingly; and (iv) When a clinician finds that the AI recommendation indicate a worse possible outcome they will disregard it and seek their own alternative approach.
Collapse
|
36
|
Krakowski I, Kim J, Cai ZR, Daneshjou R, Lapins J, Eriksson H, Lykou A, Linos E. Human-AI interaction in skin cancer diagnosis: a systematic review and meta-analysis. NPJ Digit Med 2024; 7:78. [PMID: 38594408 PMCID: PMC11004168 DOI: 10.1038/s41746-024-01031-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2023] [Accepted: 02/05/2024] [Indexed: 04/11/2024] Open
Abstract
The development of diagnostic tools for skin cancer based on artificial intelligence (AI) is increasing rapidly and will likely soon be widely implemented in clinical use. Even though the performance of these algorithms is promising in theory, there is limited evidence on the impact of AI assistance on human diagnostic decisions. Therefore, the aim of this systematic review and meta-analysis was to study the effect of AI assistance on the accuracy of skin cancer diagnosis. We searched PubMed, Embase, IEE Xplore, Scopus and conference proceedings for articles from 1/1/2017 to 11/8/2022. We included studies comparing the performance of clinicians diagnosing at least one skin cancer with and without deep learning-based AI assistance. Summary estimates of sensitivity and specificity of diagnostic accuracy with versus without AI assistance were computed using a bivariate random effects model. We identified 2983 studies, of which ten were eligible for meta-analysis. For clinicians without AI assistance, pooled sensitivity was 74.8% (95% CI 68.6-80.1) and specificity was 81.5% (95% CI 73.9-87.3). For AI-assisted clinicians, the overall sensitivity was 81.1% (95% CI 74.4-86.5) and specificity was 86.1% (95% CI 79.2-90.9). AI benefitted medical professionals of all experience levels in subgroup analyses, with the largest improvement among non-dermatologists. No publication bias was detected, and sensitivity analysis revealed that the findings were robust. AI in the hands of clinicians has the potential to improve diagnostic accuracy in skin cancer diagnosis. Given that most studies were conducted in experimental settings, we encourage future studies to further investigate these potential benefits in real-life settings.
Collapse
Affiliation(s)
- Isabelle Krakowski
- Center for Digital Health, Stanford University School of Medicine, Stanford, CA, USA
- Department of Oncology and Pathology, Karolinska Institutet, Stockholm, Sweden
- Department of Dermatology, Stanford, Stanford University, Stanford, CA, USA
| | - Jiyeong Kim
- Center for Digital Health, Stanford University School of Medicine, Stanford, CA, USA
- Department of Dermatology, Stanford, Stanford University, Stanford, CA, USA
| | - Zhuo Ran Cai
- Center for Digital Health, Stanford University School of Medicine, Stanford, CA, USA
- Department of Dermatology, Stanford, Stanford University, Stanford, CA, USA
| | - Roxana Daneshjou
- Department of Dermatology, Department of Biomedical Data Science, Stanford School of Medicine, Stanford, CA, USA
| | - Jan Lapins
- Department of Dermatology, Theme Inflammation, Karolinska University Hospital, Stockholm, Sweden
| | - Hanna Eriksson
- Department of Oncology and Pathology, Karolinska Institutet, Stockholm, Sweden
- Theme Cancer, Unit of Head-Neck-, Lung- and Skin Cancer, Skin Cancer Center, Karolinska University Hospital, Stockholm, Sweden
| | - Anastasia Lykou
- Department of Education, University of Nicosia, Nicosia, Cyprus
| | - Eleni Linos
- Center for Digital Health, Stanford University School of Medicine, Stanford, CA, USA.
- Department of Dermatology, Stanford, Stanford University, Stanford, CA, USA.
| |
Collapse
|
37
|
Primiero CA, Betz-Stablein B, Ascott N, D’Alessandro B, Gaborit S, Fricker P, Goldsteen A, González-Villà S, Lee K, Nazari S, Nguyen H, Ntouskos V, Pahde F, Pataki BE, Quintana J, Puig S, Rezze GG, Garcia R, Soyer HP, Malvehy J. A protocol for annotation of total body photography for machine learning to analyze skin phenotype and lesion classification. Front Med (Lausanne) 2024; 11:1380984. [PMID: 38654834 PMCID: PMC11035726 DOI: 10.3389/fmed.2024.1380984] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2024] [Accepted: 03/25/2024] [Indexed: 04/26/2024] Open
Abstract
Introduction Artificial Intelligence (AI) has proven effective in classifying skin cancers using dermoscopy images. In experimental settings, algorithms have outperformed expert dermatologists in classifying melanoma and keratinocyte cancers. However, clinical application is limited when algorithms are presented with 'untrained' or out-of-distribution lesion categories, often misclassifying benign lesions as malignant, or misclassifying malignant lesions as benign. Another limitation often raised is the lack of clinical context (e.g., medical history) used as input for the AI decision process. The increasing use of Total Body Photography (TBP) in clinical examinations presents new opportunities for AI to perform holistic analysis of the whole patient, rather than a single lesion. Currently there is a lack of existing literature or standards for image annotation of TBP, or on preserving patient privacy during the machine learning process. Methods This protocol describes the methods for the acquisition of patient data, including TBP, medical history, and genetic risk factors, to create a comprehensive dataset for machine learning. 500 patients of various risk profiles will be recruited from two clinical sites (Australia and Spain), to undergo temporal total body imaging, complete surveys on sun behaviors and medical history, and provide a DNA sample. This patient-level metadata is applied to image datasets using DICOM labels. Anonymization and masking methods are applied to preserve patient privacy. A two-step annotation process is followed to label skin images for lesion detection and classification using deep learning models. Skin phenotype characteristics are extracted from images, including innate and facultative skin color, nevi distribution, and UV damage. Several algorithms will be developed relating to skin lesion detection, segmentation and classification, 3D mapping, change detection, and risk profiling. Simultaneously, explainable AI (XAI) methods will be incorporated to foster clinician and patient trust. Additionally, a publicly released dataset of anonymized annotated TBP images will be released for an international challenge to advance the development of new algorithms using this type of data. Conclusion The anticipated results from this protocol are validated AI-based tools to provide holistic risk assessment for individual lesions, and risk stratification of patients to assist clinicians in monitoring for skin cancer.
Collapse
Affiliation(s)
- Clare A. Primiero
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica—IDIBAPS, Barcelona, Spain
- Frazer Institute, The University of Queensland, Dermatology Research Centre, Brisbane, QLD, Australia
| | - Brigid Betz-Stablein
- Frazer Institute, The University of Queensland, Dermatology Research Centre, Brisbane, QLD, Australia
| | | | | | | | - Paul Fricker
- Torus Actions & Belle.ai, Ramonville-Saint-Agne, France
| | | | | | - Katie Lee
- Frazer Institute, The University of Queensland, Dermatology Research Centre, Brisbane, QLD, Australia
| | - Sana Nazari
- Computer Vision and Robotics Group, University of Girona, Girona, Spain
| | - Hang Nguyen
- Torus Actions & Belle.ai, Ramonville-Saint-Agne, France
| | - Valsamis Ntouskos
- Remote Sensing Lab, National Technical University of Athens, Athens, Greece
| | | | - Balázs E. Pataki
- HUN-REN Institute for Computer Science and Control, Budapest, Hungary
| | | | - Susana Puig
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica—IDIBAPS, Barcelona, Spain
- Medicine Department, University of Barcelona, Barcelona, Spain
- CIBER de Enfermedades raras, Instituto de Salud Carlos III, Barcelona, Spain
| | - Gisele G. Rezze
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica—IDIBAPS, Barcelona, Spain
| | - Rafael Garcia
- Computer Vision and Robotics Group, University of Girona, Girona, Spain
| | - H. Peter Soyer
- Frazer Institute, The University of Queensland, Dermatology Research Centre, Brisbane, QLD, Australia
- Dermatology Department, Princess Alexandra Hospital, Brisbane, QLD, Australia
| | - Josep Malvehy
- Dermatology Department, Hospital Clinic and Fundació Clínic per la Recerca Biomèdica—IDIBAPS, Barcelona, Spain
- Medicine Department, University of Barcelona, Barcelona, Spain
- CIBER de Enfermedades raras, Instituto de Salud Carlos III, Barcelona, Spain
| |
Collapse
|
38
|
Miller I, Rosic N, Stapelberg M, Hudson J, Coxon P, Furness J, Walsh J, Climstein M. Performance of Commercial Dermatoscopic Systems That Incorporate Artificial Intelligence for the Identification of Melanoma in General Practice: A Systematic Review. Cancers (Basel) 2024; 16:1443. [PMID: 38611119 PMCID: PMC11011068 DOI: 10.3390/cancers16071443] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2024] [Revised: 04/03/2024] [Accepted: 04/04/2024] [Indexed: 04/14/2024] Open
Abstract
BACKGROUND Cutaneous melanoma remains an increasing global public health burden, particularly in fair-skinned populations. Advancing technologies, particularly artificial intelligence (AI), may provide an additional tool for clinicians to help detect malignancies with a more accurate success rate. This systematic review aimed to report the performance metrics of commercially available convolutional neural networks (CNNs) tasked with detecting MM. METHODS A systematic literature search was performed using CINAHL, Medline, Scopus, ScienceDirect and Web of Science databases. RESULTS A total of 16 articles reporting MM were included in this review. The combined number of melanomas detected was 1160, and non-melanoma lesions were 33,010. The performance of market-approved technology and clinician performance for classifying melanoma was highly heterogeneous, with sensitivity ranging from 16.4 to 100.0%, specificity between 40.0 and 98.3% and accuracy between 44.0 and 92.0%. Less heterogeneity was observed when clinicians worked in unison with AI, with sensitivity ranging between 83.3 and 100.0%, specificity between 83.7 and 87.3%, and accuracy between 86.4 and 86.9%. CONCLUSION Instead of focusing on the performance of AI versus clinicians for classifying melanoma, more consistent performance has been obtained when clinicians' work is supported by AI, facilitating management decisions and improving health outcomes.
Collapse
Affiliation(s)
- Ian Miller
- Aquatic Based Research, Southern Cross University, Bilinga, QLD 4225, Australia; (I.M.); (N.R.)
- Faculty of Health, Southern Cross University, Bilinga, QLD 4225, Australia (P.C.)
- Specialist Suite, John Flynn Hospital, Tugun, QLD 4224, Australia
| | - Nedeljka Rosic
- Aquatic Based Research, Southern Cross University, Bilinga, QLD 4225, Australia; (I.M.); (N.R.)
- Faculty of Health, Southern Cross University, Bilinga, QLD 4225, Australia (P.C.)
| | - Michael Stapelberg
- Aquatic Based Research, Southern Cross University, Bilinga, QLD 4225, Australia; (I.M.); (N.R.)
- Faculty of Health, Southern Cross University, Bilinga, QLD 4225, Australia (P.C.)
- Specialist Suite, John Flynn Hospital, Tugun, QLD 4224, Australia
| | - Jeremy Hudson
- Faculty of Health, Southern Cross University, Bilinga, QLD 4225, Australia (P.C.)
- North Queensland Skin Centre, Townsville, QLD 4810, Australia
| | - Paul Coxon
- Faculty of Health, Southern Cross University, Bilinga, QLD 4225, Australia (P.C.)
- North Queensland Skin Centre, Townsville, QLD 4810, Australia
| | - James Furness
- Water Based Research Unit, Bond University, Robina, QLD 4226, Australia;
| | - Joe Walsh
- Sport Science Institute, Sydney, NSW 2000, Australia;
- AI Consulting Group, Sydney, NSW 2000, Australia
| | - Mike Climstein
- Aquatic Based Research, Southern Cross University, Bilinga, QLD 4225, Australia; (I.M.); (N.R.)
- Faculty of Health, Southern Cross University, Bilinga, QLD 4225, Australia (P.C.)
- Physical Activity, Lifestyle, Ageing and Wellbeing Faculty Research Group, University of Sydney, Sydney, NSW 2050, Australia
| |
Collapse
|
39
|
Zhang P, Gao C, Huang Y, Chen X, Pan Z, Wang L, Dong D, Li S, Qi X. Artificial intelligence in liver imaging: methods and applications. Hepatol Int 2024; 18:422-434. [PMID: 38376649 DOI: 10.1007/s12072-023-10630-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/15/2023] [Accepted: 12/18/2023] [Indexed: 02/21/2024]
Abstract
Liver disease is regarded as one of the major health threats to humans. Radiographic assessments hold promise in terms of addressing the current demands for precisely diagnosing and treating liver diseases, and artificial intelligence (AI), which excels at automatically making quantitative assessments of complex medical image characteristics, has made great strides regarding the qualitative interpretation of medical imaging by clinicians. Here, we review the current state of medical-imaging-based AI methodologies and their applications concerning the management of liver diseases. We summarize the representative AI methodologies in liver imaging with focusing on deep learning, and illustrate their promising clinical applications across the spectrum of precise liver disease detection, diagnosis and treatment. We also address the current challenges and future perspectives of AI in liver imaging, with an emphasis on feature interpretability, multimodal data integration and multicenter study. Taken together, it is revealed that AI methodologies, together with the large volume of available medical image data, might impact the future of liver disease care.
Collapse
Affiliation(s)
- Peng Zhang
- Institute for TCM-X, MOE Key Laboratory of Bioinformatics, Bioinformatics Division, BNRIST, Department of Automation, Tsinghua University, Beijing, China
| | - Chaofei Gao
- Institute for TCM-X, MOE Key Laboratory of Bioinformatics, Bioinformatics Division, BNRIST, Department of Automation, Tsinghua University, Beijing, China
| | - Yifei Huang
- Department of Gastroenterology, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xiangyi Chen
- Institute for TCM-X, MOE Key Laboratory of Bioinformatics, Bioinformatics Division, BNRIST, Department of Automation, Tsinghua University, Beijing, China
| | - Zhuoshi Pan
- Institute for TCM-X, MOE Key Laboratory of Bioinformatics, Bioinformatics Division, BNRIST, Department of Automation, Tsinghua University, Beijing, China
| | - Lan Wang
- Institute for TCM-X, MOE Key Laboratory of Bioinformatics, Bioinformatics Division, BNRIST, Department of Automation, Tsinghua University, Beijing, China
| | - Di Dong
- CAS Key Laboratory of Molecular Imaging, Beijing Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, 100190, China
| | - Shao Li
- Institute for TCM-X, MOE Key Laboratory of Bioinformatics, Bioinformatics Division, BNRIST, Department of Automation, Tsinghua University, Beijing, China.
| | - Xiaolong Qi
- Center of Portal Hypertension, Department of Radiology, Zhongda Hospital, Medical School, Nurturing Center of Jiangsu Province for State Laboratory of AI Imaging & Interventional Radiology, Southeast University, Nanjing, China.
| |
Collapse
|
40
|
Zhong Y, Piao Y, Tan B, Liu J. A multi-task fusion model based on a residual-Multi-layer perceptron network for mammographic breast cancer screening. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 247:108101. [PMID: 38432087 DOI: 10.1016/j.cmpb.2024.108101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/20/2023] [Revised: 01/13/2024] [Accepted: 02/23/2024] [Indexed: 03/05/2024]
Abstract
BACKGROUND AND OBJECTIVE Deep learning approaches are being increasingly applied for medical computer-aided diagnosis (CAD). However, these methods generally target only specific image-processing tasks, such as lesion segmentation or benign state prediction. For the breast cancer screening task, single feature extraction models are generally used, which directly extract only those potential features from the input mammogram that are relevant to the target task. This can lead to the neglect of other important morphological features of the lesion as well as other auxiliary information from the internal breast tissue. To obtain more comprehensive and objective diagnostic results, in this study, we developed a multi-task fusion model that combines multiple specific tasks for CAD of mammograms. METHODS We first trained a set of separate, task-specific models, including a density classification model, a mass segmentation model, and a lesion benignity-malignancy classification model, and then developed a multi-task fusion model that incorporates all of the mammographic features from these different tasks to yield comprehensive and refined prediction results for breast cancer diagnosis. RESULTS The experimental results showed that our proposed multi-task fusion model outperformed other related state-of-the-art models in both breast cancer screening tasks in the publicly available datasets CBIS-DDSM and INbreast, achieving a competitive screening performance with area-under-the-curve scores of 0.92 and 0.95, respectively. CONCLUSIONS Our model not only allows an overall assessment of lesion types in mammography but also provides intermediate results related to radiological features and potential cancer risk factors, indicating its potential to offer comprehensive workflow support to radiologists.
Collapse
Affiliation(s)
- Yutong Zhong
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun 130022, PR China
| | - Yan Piao
- School of Electronic Information Engineering, Changchun University of Science and Technology, Changchun 130022, PR China.
| | - Baolin Tan
- Technology Co. LTD, Shenzhen 518000, PR China
| | - Jingxin Liu
- Department of Radiology, China-Japan Union Hospital, Jilin University, Changchun 130033, PR China
| |
Collapse
|
41
|
Li T, Ma X, Li Z, Yu N, Song J, Ma Z, Ying H, Zhou B, Huang J, Wu L, Long X. Facial adult female acne in China: An analysis based on artificial intelligence over one million. Skin Res Technol 2024; 30:e13693. [PMID: 38572573 PMCID: PMC10993051 DOI: 10.1111/srt.13693] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2024] [Accepted: 02/19/2024] [Indexed: 04/05/2024]
Abstract
BACKGROUND To further clarify the acne profile of Chinese adult women, we included 1,156,703 adult women. An artificial intelligence algorithm was used to analyze images taken by high-resolution mobile phones to further explore acne levels in Chinese adult women. METHOD In this study, we assessed the severity of acne by evaluating patients' selfies through a smartphone application. Furthermore, we gathered basic user information through a questionnaire, including details such as age, gender, skin sensitivity, and dietary habits. RESULTS This study showed a gradual decrease in acne severity from the age of 25 years. A trough was reached between the ages of 40 and 44, followed by a gradual increase in acne severity. In terms of skin problems and acne severity, we have found that oily skin, hypersensitive skin, frequent makeup application and unhealthy dietary habits can affect the severity of acne. For environment and acne severity, we observed that developed city levels, cold seasons and high altitude and strong radiation affect acne severity in adult women. For the results of the AI analyses, the severity of blackheads, pores, dark circles and skin roughness were positively associated with acne severity in adult women. CONCLUSIONS AI analysis of high-res phone images in Chinese adult women reveals acne severity trends. Severity decreases after 25, hits a low at 40-44, then gradually rises. Skin type, sensitivity, makeup, diet, urbanization, seasons, altitude, and radiation impact acne. Blackheads, pores, dark circles, and skin roughness are linked to acne severity. These findings inform personalized skincare and public health strategies for adult women.
Collapse
Affiliation(s)
- Tian‐Hao Li
- Department of Plastic and Reconstructive SurgeryPeking Union Medical College HospitalChinese Academy of Medical Sciences and Peking Union Medical CollegeBeijingChina
| | - Xu‐Da Ma
- Department of Plastic and Reconstructive SurgeryPeking Union Medical College HospitalChinese Academy of Medical Sciences and Peking Union Medical CollegeBeijingChina
| | - Zi‐Ming Li
- Department of Plastic and Reconstructive SurgeryPeking Union Medical College HospitalChinese Academy of Medical Sciences and Peking Union Medical CollegeBeijingChina
| | - Nan‐Ze Yu
- Department of Plastic and Reconstructive SurgeryPeking Union Medical College HospitalChinese Academy of Medical Sciences and Peking Union Medical CollegeBeijingChina
| | - Jin‐Yan Song
- Hangzhou C2H4 Internet Technology Co., LtdHangzhou C2H4 Internet Technology Co., LtdHangzhouChina
| | - Zi‐Tao Ma
- Hangzhou C2H4 Internet Technology Co., LtdHangzhou C2H4 Internet Technology Co., LtdHangzhouChina
| | - Han‐Ting Ying
- Hangzhou C2H4 Internet Technology Co., LtdHangzhou C2H4 Internet Technology Co., LtdHangzhouChina
| | - Beibei Zhou
- Hangzhou C2H4 Internet Technology Co., LtdHangzhou C2H4 Internet Technology Co., LtdHangzhouChina
| | - Jiu‐Zuo Huang
- Department of Plastic and Reconstructive SurgeryPeking Union Medical College HospitalChinese Academy of Medical Sciences and Peking Union Medical CollegeBeijingChina
| | - Liang Wu
- Hangzhou C2H4 Internet Technology Co., LtdHangzhou C2H4 Internet Technology Co., LtdHangzhouChina
| | - Xiao Long
- Department of Plastic and Reconstructive SurgeryPeking Union Medical College HospitalChinese Academy of Medical Sciences and Peking Union Medical CollegeBeijingChina
| |
Collapse
|
42
|
Carmichael J, Costanza E, Blandford A, Struyven R, Keane PA, Balaskas K. Diagnostic decisions of specialist optometrists exposed to ambiguous deep-learning outputs. Sci Rep 2024; 14:6775. [PMID: 38514657 PMCID: PMC10958016 DOI: 10.1038/s41598-024-55410-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Accepted: 02/23/2024] [Indexed: 03/23/2024] Open
Abstract
Artificial intelligence (AI) has great potential in ophthalmology. We investigated how ambiguous outputs from an AI diagnostic support system (AI-DSS) affected diagnostic responses from optometrists when assessing cases of suspected retinal disease. Thirty optometrists (15 more experienced, 15 less) assessed 30 clinical cases. For ten, participants saw an optical coherence tomography (OCT) scan, basic clinical information and retinal photography ('no AI'). For another ten, they were also given AI-generated OCT-based probabilistic diagnoses ('AI diagnosis'); and for ten, both AI-diagnosis and AI-generated OCT segmentations ('AI diagnosis + segmentation') were provided. Cases were matched across the three types of presentation and were selected to include 40% ambiguous and 20% incorrect AI outputs. Optometrist diagnostic agreement with the predefined reference standard was lowest for 'AI diagnosis + segmentation' (204/300, 68%) compared to 'AI diagnosis' (224/300, 75% p = 0.010), and 'no Al' (242/300, 81%, p = < 0.001). Agreement with AI diagnosis consistent with the reference standard decreased (174/210 vs 199/210, p = 0.003), but participants trusted the AI more (p = 0.029) with segmentations. Practitioner experience did not affect diagnostic responses (p = 0.24). More experienced participants were more confident (p = 0.012) and trusted the AI less (p = 0.038). Our findings also highlight issues around reference standard definition.
Collapse
Affiliation(s)
- Josie Carmichael
- University College London Interaction Centre (UCLIC), UCL, London, UK.
- Institute of Ophthalmology, NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL, London, UK.
| | - Enrico Costanza
- University College London Interaction Centre (UCLIC), UCL, London, UK
| | - Ann Blandford
- University College London Interaction Centre (UCLIC), UCL, London, UK
| | - Robbert Struyven
- Institute of Ophthalmology, NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL, London, UK
| | - Pearse A Keane
- Institute of Ophthalmology, NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL, London, UK
| | - Konstantinos Balaskas
- Institute of Ophthalmology, NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust and UCL, London, UK
| |
Collapse
|
43
|
Wei ML, Tada M, So A, Torres R. Artificial intelligence and skin cancer. Front Med (Lausanne) 2024; 11:1331895. [PMID: 38566925 PMCID: PMC10985205 DOI: 10.3389/fmed.2024.1331895] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Accepted: 02/26/2024] [Indexed: 04/04/2024] Open
Abstract
Artificial intelligence is poised to rapidly reshape many fields, including that of skin cancer screening and diagnosis, both as a disruptive and assistive technology. Together with the collection and availability of large medical data sets, artificial intelligence will become a powerful tool that can be leveraged by physicians in their diagnoses and treatment plans for patients. This comprehensive review focuses on current progress toward AI applications for patients, primary care providers, dermatologists, and dermatopathologists, explores the diverse applications of image and molecular processing for skin cancer, and highlights AI's potential for patient self-screening and improving diagnostic accuracy for non-dermatologists. We additionally delve into the challenges and barriers to clinical implementation, paths forward for implementation and areas of active research.
Collapse
Affiliation(s)
- Maria L. Wei
- Department of Dermatology, University of California, San Francisco, San Francisco, CA, United States
- Dermatology Service, San Francisco VA Health Care System, San Francisco, CA, United States
| | - Mikio Tada
- Institute for Neurodegenerative Diseases, University of California, San Francisco, San Francisco, CA, United States
| | - Alexandra So
- School of Medicine, University of California, San Francisco, San Francisco, CA, United States
| | - Rodrigo Torres
- Dermatology Service, San Francisco VA Health Care System, San Francisco, CA, United States
| |
Collapse
|
44
|
Rozario SY, Sarkar M, Farlie MK, Lazarus MD. Responding to the healthcare workforce shortage: A scoping review exploring anatomical pathologists' professional identities over time. ANATOMICAL SCIENCES EDUCATION 2024; 17:351-365. [PMID: 36748328 DOI: 10.1002/ase.2260] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/06/2022] [Revised: 01/16/2023] [Accepted: 02/06/2023] [Indexed: 06/18/2023]
Abstract
Anatomical pathology (AP) is an anatomy-centric medical specialty devoted to tissue-based diagnosis of disease. The field faces a current and predicted workforce shortage, likely increasing diagnostic wait times and delaying patient access to urgent treatment. A lack of AP exposure is proposed to preclude recruitment to the field, as medical students are afforded only a limited understanding of who a pathologist is and what they do (their professional identity/PI and role). Anatomical sciences educators may be well placed to increase student understanding of anatomical pathologists' PI features, but until features of anatomical pathologists' PI are understood, recommendations for anatomy educators are premature. Thus, this scoping review asked: "What are the professional identity features of anatomical pathologists reported in the literature, and how have these changed over time?" A six-stage scoping review was performed. Medline and PubMed, Global Health, and Embase were used to identify relevant studies (n = 74). Team-based framework analysis identified that features of anatomical pathologists' professional identity encompass five overarching themes: professional practice, views about the role, training and education, personal implications, and technology. Technology was identified as an important theme of anatomical pathologists' PI, as it intersected with many other PI feature themes, including diagnosis and collaboration. This review found that pathologists may sometimes perceive professional competition with technology, such as artificial intelligence. These findings suggest unique opportunities for integrating AP-specific PI features into anatomy teaching, which may foster student interest in AP, and potentially increase recruitment into the field.
Collapse
Affiliation(s)
- Shemona Y Rozario
- Centre for Human Anatomy Education (CHAE), Department of Anatomy and Developmental Biology, Biomedical Discovery Institute, Faculty of Medicine, Nursing and Health Sciences, Monash University, Clayton, Victoria, Australia
| | - Mahbub Sarkar
- Monash Centre for Scholarship in Health Education (MCSHE), Faculty of Medicine, Nursing and Health Sciences, Monash University, Clayton, Victoria, Australia
| | - Melanie K Farlie
- Monash Centre for Scholarship in Health Education (MCSHE), Faculty of Medicine, Nursing and Health Sciences, Monash University, Clayton, Victoria, Australia
- Department of Physiotherapy, Faculty of Medicine Nursing and Health Sciences, Monash University, Clayton, Victoria, Australia
| | - Michelle D Lazarus
- Centre for Human Anatomy Education (CHAE), Department of Anatomy and Developmental Biology, Biomedical Discovery Institute, Faculty of Medicine, Nursing and Health Sciences, Monash University, Clayton, Victoria, Australia
- Monash Centre for Scholarship in Health Education (MCSHE), Faculty of Medicine, Nursing and Health Sciences, Monash University, Clayton, Victoria, Australia
| |
Collapse
|
45
|
Hartmann T, Passauer J, Hartmann J, Schmidberger L, Kneilling M, Volc S. Basic principles of artificial intelligence in dermatology explained using melanoma. J Dtsch Dermatol Ges 2024; 22:339-347. [PMID: 38361141 DOI: 10.1111/ddg.15322] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Accepted: 11/04/2023] [Indexed: 02/17/2024]
Abstract
The use of artificial intelligence (AI) continues to establish itself in the most diverse areas of medicine at an increasingly fast pace. Nevertheless, many healthcare professionals lack the basic technical understanding of how this technology works, which severely limits its application in clinical settings and research. Thus, we would like to discuss the functioning and classification of AI using melanoma as an example in this review to build an understanding of the technology behind AI. For this purpose, elaborate illustrations are used that quickly reveal the technology involved. Previous reviews tend to focus on the potential applications of AI, thereby missing the opportunity to develop a deeper understanding of the subject matter that is so important for clinical application. Malignant melanoma has become a significant burden for healthcare systems. If discovered early, a better prognosis can be expected, which is why skin cancer screening has become increasingly popular and is supported by health insurance. The number of experts remains finite, reducing their availability and leading to longer waiting times. Therefore, innovative ideas need to be implemented to provide the necessary care. Thus, machine learning offers the ability to recognize melanomas from images at a level comparable to experienced dermatologists under optimized conditions.
Collapse
Affiliation(s)
- Tim Hartmann
- Department of Dermatology, University hospital Tübingen, Tübingen, Germany
| | - Johannes Passauer
- Department of Dermatology, University hospital Tübingen, Tübingen, Germany
| | | | - Laura Schmidberger
- Department of Dermatology, University hospital Tübingen, Tübingen, Germany
| | - Manfred Kneilling
- Department of Dermatology, University hospital Tübingen, Tübingen, Germany
- Werner Siemens Imaging Center, Department of Preclinical Imaging and Radiopharmacy, Eberhard Karls University, Tübingen, Germany
- Cluster of Excellence iFIT (EXC 2180) "Image-Guided and Functionally Instructed Tumor Therapies", Eberhard Karls University, Tübingen, Germany
| | - Sebastian Volc
- Department of Dermatology, University hospital Tübingen, Tübingen, Germany
| |
Collapse
|
46
|
Hartmann T, Passauer J, Hartmann J, Schmidberger L, Kneilling M, Volc S. Grundprinzipien der künstlichen Intelligenz in der Dermatologie erklärt am Beispiel des Melanoms. J Dtsch Dermatol Ges 2024; 22:339-349. [PMID: 38450927 DOI: 10.1111/ddg.15322_g] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Accepted: 11/04/2023] [Indexed: 03/08/2024]
Abstract
ZusammenfassungDer Einsatz von künstlicher Intelligenz (KI) setzt sich in den verschiedensten Bereichen der Medizin immer schneller durch. Dennoch fehlt vielen medizinischen Kollegen das technische Grundverständnis für die Funktionsweise dieser Technologie, was ihre Anwendung in Klinik und Forschung stark einschränkt. Daher möchten wir in dieser Übersichtsarbeit die Funktionsweise und Klassifizierung der KI am Beispiel des Melanoms erörtern, um ein Verständnis für die Technologie hinter der KI zu schaffen. Dazu werden ausführliche Illustrationen verwendet, die die Technologie schnell erklären. Bisherige Übersichten konzentrieren sich eher auf die potenziellen Anwendungen der KI und verpassen die Gelegenheit, ein tieferes Verständnis für die Materie herauszuarbeiten, das für die klinische Anwendung so wichtig ist. Das maligne Melanom ist zu einer erheblichen Belastung für die Gesundheitssysteme geworden. Bei frühzeitiger Entdeckung ist eine bessere Prognose zu erwarten, weshalb das Hautkrebs‐Screening immer populärer und von den Krankenkassen unterstützt wird. Die Zahl der Fachärzte ist jedoch begrenzt, was ihre Verfügbarkeit einschränkt und zu längeren Wartezeiten führt. Daher müssen innovative Ideen umgesetzt werden, um die notwendige Versorgung zu gewährleisten. Das maschinelle Lernen bietet die Möglichkeit, Melanome auf Bildern zu erkennen, und zwar auf einem Niveau, das mit dem von erfahrenen Dermatologen – unter optimierten Bedingungen – vergleichbar ist.
Collapse
Affiliation(s)
- Tim Hartmann
- Hautklinik, Universitätsklinik, Eberhard Karls Universität, Tübingen
| | - Johannes Passauer
- Hautklinik, Universitätsklinik, Eberhard Karls Universität, Tübingen
| | | | | | - Manfred Kneilling
- Hautklinik, Universitätsklinik, Eberhard Karls Universität, Tübingen
- Werner Siemens Imaging Center, Department of Preclinical Imaging and Radiopharmacy, Eberhard Karls University, Tübingen
- Cluster of Excellence iFIT (EXC 2180) "Image-Guided and Functionally Instructed Tumor Therapies", Eberhard Karls Universität, Tübingen
| | - Sebastian Volc
- Hautklinik, Universitätsklinik, Eberhard Karls Universität, Tübingen
| |
Collapse
|
47
|
Li Z, Zhang N, Gong H, Qiu R, Zhang W. SG-MIAN: Self-guided multiple information aggregation network for image-level weakly supervised skin lesion segmentation. Comput Biol Med 2024; 170:107988. [PMID: 38232452 DOI: 10.1016/j.compbiomed.2024.107988] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2023] [Revised: 12/11/2023] [Accepted: 01/13/2024] [Indexed: 01/19/2024]
Abstract
Nowadays, skin disease is becoming one of the most malignant diseases that threaten people's health. Computer aided diagnosis based on deep learning has become a widely used technology to assist medical professionals in diagnosis, and segmentation of lesion areas is one of the most important steps in it. However, traditional medical image segmentation methods rely on numerous pixel-level labels for fully supervised training, and such labeling process is time-consuming and requires professional competence. In order to reduce the costs of pixel-level labeling, we proposed a method only using image-level label to segment skin lesion areas. Due to the lack of lesion's spatial and intensity information in image-level labels, and the wide distribution range of irregular shape and different texture on skin lesions, the algorithm must pay great attention to the automatic lesion localization and perception of lesion boundary. In this paper, we proposed a Self-Guided Multiple Information Aggregation Network (SG-MIAN). Our backbone network MIAN utilizes the Multiple Spatial Perceptron (MSP) solely using classification information as guidance to discriminate the key classification features of lesion areas, and thereby performing more accurate localization and activation of lesion areas. Additionally, adjunct to MSP, we also proposed an Auxiliary Activation Structure (AAS) and two auxiliary loss functions to further self-guided boundary correction, achieving the goal of accurate boundary activation. To verify the effectiveness of the proposed method, we conducted extensive experiments using the HAM10000 dataset and the PH2dataset, which demonstrated superior performance compared to most existing weakly supervised segmentation methods.
Collapse
Affiliation(s)
- Zhixun Li
- School of Mathematics and Computer Sciences, Nanchang University, Nanchang, China
| | - Nan Zhang
- School of Mathematics and Computer Sciences, Nanchang University, Nanchang, China
| | - Huiling Gong
- School of Mathematics and Computer Sciences, Nanchang University, Nanchang, China
| | - Ruiyun Qiu
- School of Mathematics and Computer Sciences, Nanchang University, Nanchang, China.
| | - Wei Zhang
- School of Mathematics and Computer Sciences, Nanchang University, Nanchang, China
| |
Collapse
|
48
|
Brancaccio G, Balato A, Malvehy J, Puig S, Argenziano G, Kittler H. Artificial Intelligence in Skin Cancer Diagnosis: A Reality Check. J Invest Dermatol 2024; 144:492-499. [PMID: 37978982 DOI: 10.1016/j.jid.2023.10.004] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2023] [Revised: 09/08/2023] [Accepted: 10/01/2023] [Indexed: 11/19/2023]
Abstract
The field of skin cancer detection offers a compelling use case for the application of artificial intelligence (AI) within the realm of image-based diagnostic medicine. Through the analysis of large datasets, AI algorithms have the capacity to classify clinical or dermoscopic images with remarkable accuracy. Although these AI-based applications can operate both autonomously and under human supervision, the best results are achieved through a collaborative approach that leverages the expertise of both AI and human experts. However, it is important to note that most studies focus on assessing the diagnostic accuracy of AI in artificial settings rather than in real-world scenarios. Consequently, the practical utility of AI-assisted diagnosis in a clinical environment is still largely unknown. Furthermore, there exists a knowledge gap concerning the optimal use cases and deployment settings for these AI systems as well as the practical challenges that may arise from widespread implementation. This review explores the advantages and limitations of AI in a variety of real-world contexts, with a specific focus on its value to consumers, general practitioners, and dermatologists.
Collapse
Affiliation(s)
| | - Anna Balato
- Dermatology Unit, University of Campania "Luigi Vanvitelli", Naples, Italy
| | - Josep Malvehy
- Melanoma Unit, Dermatology Department, Hospital Clínic de Barcelona, Instituto de Investigaciones Biomédicas August Pi i Sunye, Universitat de Barcelona, Barcelona, Spain; Centro de Investigación Biomédica en Red de Enfermedades Raras (CIBERER), Instituto de Salud Carlos III, Barcelona, Spain
| | - Susana Puig
- Melanoma Unit, Dermatology Department, Hospital Clínic de Barcelona, Instituto de Investigaciones Biomédicas August Pi i Sunye, Universitat de Barcelona, Barcelona, Spain; Centro de Investigación Biomédica en Red de Enfermedades Raras (CIBERER), Instituto de Salud Carlos III, Barcelona, Spain
| | | | - Harald Kittler
- Department of Dermatology, Medical University of Vienna, Vienna, Austria
| |
Collapse
|
49
|
Subramanian HV, Canfield C, Shank DB. Designing explainable AI to improve human-AI team performance: A medical stakeholder-driven scoping review. Artif Intell Med 2024; 149:102780. [PMID: 38462282 DOI: 10.1016/j.artmed.2024.102780] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2023] [Revised: 12/20/2023] [Accepted: 01/14/2024] [Indexed: 03/12/2024]
Abstract
The rise of complex AI systems in healthcare and other sectors has led to a growing area of research called Explainable AI (XAI) designed to increase transparency. In this area, quantitative and qualitative studies focus on improving user trust and task performance by providing system- and prediction-level XAI features. We analyze stakeholder engagement events (interviews and workshops) on the use of AI for kidney transplantation. From this we identify themes which we use to frame a scoping literature review on current XAI features. The stakeholder engagement process lasted over nine months covering three stakeholder group's workflows, determining where AI could intervene and assessing a mock XAI decision support system. Based on the stakeholder engagement, we identify four major themes relevant to designing XAI systems - 1) use of AI predictions, 2) information included in AI predictions, 3) personalization of AI predictions for individual differences, and 4) customizing AI predictions for specific cases. Using these themes, our scoping literature review finds that providing AI predictions before, during, or after decision-making could be beneficial depending on the complexity of the stakeholder's task. Additionally, expert stakeholders like surgeons prefer minimal to no XAI features, AI prediction, and uncertainty estimates for easy use cases. However, almost all stakeholders prefer to have optional XAI features to review when needed, especially in hard-to-predict cases. The literature also suggests that providing both system- and prediction-level information is necessary to build the user's mental model of the system appropriately. Although XAI features improve users' trust in the system, human-AI team performance is not always enhanced. Overall, stakeholders prefer to have agency over the XAI interface to control the level of information based on their needs and task complexity. We conclude with suggestions for future research, especially on customizing XAI features based on preferences and tasks.
Collapse
Affiliation(s)
- Harishankar V Subramanian
- Engineering Management & Systems Engineering, Missouri University of Science and Technology, 600 W 14(th) Street, Rolla, MO 65409, United States of America
| | - Casey Canfield
- Engineering Management & Systems Engineering, Missouri University of Science and Technology, 600 W 14(th) Street, Rolla, MO 65409, United States of America.
| | - Daniel B Shank
- Psychological Science, Missouri University of Science and Technology, 500 W 14(th) Street, Rolla, MO 65409, United States of America
| |
Collapse
|
50
|
Liopyris K, Navarrete-Dechent C, Marchetti MA, Rotemberg V, Apalla Z, Argenziano G, Blum A, Braun RP, Carrera C, Codella NCF, Combalia M, Dusza SW, Gutman DA, Helba B, Hofmann-Wellenhof R, Jaimes N, Kittler H, Kose K, Lallas A, Longo C, Malvehy J, Menzies S, Nelson KC, Paoli J, Puig S, Rabinovitz HS, Rishpon A, Russo T, Scope A, Soyer HP, Stein JA, Stolz W, Sgouros D, Stratigos AJ, Swanson DL, Thomas L, Tschandl P, Zalaudek I, Weber J, Halpern AC, Marghoob AA. Expert Agreement on the Presence and Spatial Localization of Melanocytic Features in Dermoscopy. J Invest Dermatol 2024; 144:531-539.e13. [PMID: 37689267 DOI: 10.1016/j.jid.2023.01.045] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2022] [Accepted: 01/19/2023] [Indexed: 09/11/2023]
Abstract
Dermoscopy aids in melanoma detection; however, agreement on dermoscopic features, including those of high clinical relevance, remains poor. In this study, we attempted to evaluate agreement among experts on exemplar images not only for the presence of melanocytic-specific features but also for spatial localization. This was a cross-sectional, multicenter, observational study. Dermoscopy images exhibiting at least 1 of 31 melanocytic-specific features were submitted by 25 world experts as exemplars. Using a web-based platform that allows for image markup of specific contrast-defined regions (superpixels), 20 expert readers annotated 248 dermoscopic images in collections of 62 images. Each collection was reviewed by five independent readers. A total of 4,507 feature observations were performed. Good-to-excellent agreement was found for 14 of 31 features (45.2%), with eight achieving excellent agreement (Gwet's AC >0.75) and seven of them being melanoma-specific features. These features were peppering/granularity (0.91), shiny white streaks (0.89), typical pigment network (0.83), blotch irregular (0.82), negative network (0.81), irregular globules (0.78), dotted vessels (0.77), and blue-whitish veil (0.76). By utilizing an exemplar dataset, a good-to-excellent agreement was found for 14 features that have previously been shown useful in discriminating nevi from melanoma. All images are public (www.isic-archive.com) and can be used for education, scientific communication, and machine learning experiments.
Collapse
Affiliation(s)
- Konstantinos Liopyris
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, USA; Department of Dermatology, Andreas Syggros Hospital of Cutaneous & Venereal Diseases, University of Athens, Athens, Greece
| | - Cristian Navarrete-Dechent
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, USA; Department of Dermatology, Facultad de Medicina, Pontificia Universidad Católica de Chile, Santiago, Chile
| | - Michael A Marchetti
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, USA
| | - Veronica Rotemberg
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, USA
| | - Zoe Apalla
- First Department of Dermatology, Aristotle University School of Medicine, Thessaloniki, Greece
| | | | - Andreas Blum
- Public, Private, and Teaching Practice of Dermatology, Konstanz, Germany
| | - Ralph P Braun
- Department of Dermatology, University Hospital Zürich, Zürich, Switzerland
| | - Cristina Carrera
- Melanoma Unit, Department of Dermatology, Hospital Clínic de Barcelona, University of Barcelona, Barcelona, Spain; Centro de Investigación Biomédica en Red de Enfermedades Raras (CIBERER), Valencia, Spain
| | - Noel C F Codella
- IBM Research AI, Thomas J. Watson Research Center, Yorktown Heights, New York, USA
| | - Marc Combalia
- Melanoma Unit, Department of Dermatology, Hospital Clínic de Barcelona, University of Barcelona, Barcelona, Spain; Centro de Investigación Biomédica en Red de Enfermedades Raras (CIBERER), Valencia, Spain
| | - Stephen W Dusza
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, USA
| | - David A Gutman
- Department of Biomedical Informatics, Emory University School of Medicine, Atlanta, Georgia, USA; Department of Neurology, Emory University School of Medicine, Atlanta, Georgia, USA
| | | | | | - Natalia Jaimes
- Dr. Phillip Frost Department of Dermatology and Cutaneous Surgery, Miller School of Medicine, University of Miami, Miami, Florida, USA; Sylvester Comprehensive Cancer Center, University of Miami, Miami, Florida, USA
| | - Harald Kittler
- Vienna Dermatologic Imaging Research Group, Department of Dermatology, Medical University of Vienna, Vienna, Austria
| | - Kivanc Kose
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, USA
| | - Aimilios Lallas
- First Department of Dermatology, Aristotle University School of Medicine, Thessaloniki, Greece
| | - Caterina Longo
- Department of Dermatology, University of Modena and Reggio Emilia, Modena, Italy; Centro Oncologico ad Alta Tecnologia Diagnostica, Azienda Unità Sanitaria Locale - IRCCS di Reggio Emilia, Reggio Emilia, Italy
| | - Josep Malvehy
- Melanoma Unit, Department of Dermatology, Hospital Clínic de Barcelona, University of Barcelona, Barcelona, Spain; Centro de Investigación Biomédica en Red de Enfermedades Raras (CIBERER), Valencia, Spain
| | - Scott Menzies
- Faculty of Medicine and Health, Sydney Medical School, The University of Sydney, Camperdown, Australia; Sydney Melanoma Diagnostic Centre, Royal Prince Alfred Hospital, Camperdown, Australia
| | - Kelly C Nelson
- MD Anderson Cancer Center, Department of Dermatology, The University of Texas, Houston, Texas, USA
| | - John Paoli
- Department of Dermatology and Venereology, Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Gothenburg, Sweden
| | - Susana Puig
- Melanoma Unit, Department of Dermatology, Hospital Clínic de Barcelona, University of Barcelona, Barcelona, Spain; Centro de Investigación Biomédica en Red de Enfermedades Raras (CIBERER), Valencia, Spain
| | - Harold S Rabinovitz
- Department of Dermatology, University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Ayelet Rishpon
- Department of Dermatology, Tel Aviv Sourasky Medical Center, Tel Aviv, Israel
| | - Teresa Russo
- Dermatology Unit, University of Campania Luigi Vanvitelli, Naples, Italy
| | - Alon Scope
- Medical Screening Institute, Chaim Sheba Medical Center, Ramat Gan, Israel; Sackler School of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - H Peter Soyer
- Dermatology Research Centre, The University of Queensland Diamantina Institute, Brisbane, Australia
| | - Jennifer A Stein
- The Ronald O. Perelman Department of Dermatology, New York University School of Medicine, New York, New York, USA
| | - Willhelm Stolz
- Department of Dermatology, Ludwig-Maximilians-Universität, Munich, Germany
| | - Dimitrios Sgouros
- Department of Dermatology, Andreas Syggros Hospital of Cutaneous & Venereal Diseases, University of Athens, Athens, Greece
| | - Alexander J Stratigos
- Department of Dermatology, Andreas Syggros Hospital of Cutaneous & Venereal Diseases, University of Athens, Athens, Greece
| | - David L Swanson
- Department of Dermatology, Mayo Clinic, Scottsdale, Arizona, USA
| | - Luc Thomas
- Department of Dermatology, Centre Hospitalier de Lyon Sud, Hospices Civils de Lyon, Université Claude Bernard Lyon 1, Pierre Bénite, France
| | - Philipp Tschandl
- Vienna Dermatologic Imaging Research Group, Department of Dermatology, Medical University of Vienna, Vienna, Austria
| | - Iris Zalaudek
- Dermatology Clinic, Maggiore Hospital, University of Trieste, Trieste, Italy
| | - Jochen Weber
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, USA
| | - Allan C Halpern
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, USA
| | - Ashfaq A Marghoob
- Dermatology Service, Department of Medicine, Memorial Sloan Kettering Cancer Center, New York, USA.
| |
Collapse
|