1
|
Dehdab R, Brendlin A, Werner S, Almansour H, Gassenmaier S, Brendel JM, Nikolaou K, Afat S. Evaluating ChatGPT-4V in chest CT diagnostics: a critical image interpretation assessment. Jpn J Radiol 2024; 42:1168-1177. [PMID: 38867035 PMCID: PMC11442562 DOI: 10.1007/s11604-024-01606-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2024] [Accepted: 05/28/2024] [Indexed: 06/14/2024]
Abstract
PURPOSE To assess the diagnostic accuracy of ChatGPT-4V in interpreting a set of four chest CT slices for each case of COVID-19, non-small cell lung cancer (NSCLC), and control cases, thereby evaluating its potential as an AI tool in radiological diagnostics. MATERIALS AND METHODS In this retrospective study, 60 CT scans from The Cancer Imaging Archive, covering COVID-19, NSCLC, and control cases were analyzed using ChatGPT-4V. A radiologist selected four CT slices from each scan for evaluation. ChatGPT-4V's interpretations were compared against the gold standard diagnoses and assessed by two radiologists. Statistical analyses focused on accuracy, sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV), along with an examination of the impact of pathology location and lobe involvement. RESULTS ChatGPT-4V showed an overall diagnostic accuracy of 56.76%. For NSCLC, sensitivity was 27.27% and specificity was 60.47%. In COVID-19 detection, sensitivity was 13.64% and specificity of 64.29%. For control cases, the sensitivity was 31.82%, with a specificity of 95.24%. The highest sensitivity (83.33%) was observed in cases involving all lung lobes. The chi-squared statistical analysis indicated significant differences in Sensitivity across categories and in relation to the location and lobar involvement of pathologies. CONCLUSION ChatGPT-4V demonstrated variable diagnostic performance in chest CT interpretation, with notable proficiency in specific scenarios. This underscores the challenges of cross-modal AI models like ChatGPT-4V in radiology, pointing toward significant areas for improvement to ensure dependability. The study emphasizes the importance of enhancing these models for broader, more reliable medical use.
Collapse
Affiliation(s)
- Reza Dehdab
- Department of Diagnostic and Interventional Radiology, Tuebingen University Hospital, Hoppe-Seyler-Straße 3, 72076, Tuebingen, Germany.
| | - Andreas Brendlin
- Department of Diagnostic and Interventional Radiology, Tuebingen University Hospital, Hoppe-Seyler-Straße 3, 72076, Tuebingen, Germany
| | - Sebastian Werner
- Department of Diagnostic and Interventional Radiology, Tuebingen University Hospital, Hoppe-Seyler-Straße 3, 72076, Tuebingen, Germany
| | - Haidara Almansour
- Department of Diagnostic and Interventional Radiology, Tuebingen University Hospital, Hoppe-Seyler-Straße 3, 72076, Tuebingen, Germany
| | - Sebastian Gassenmaier
- Department of Diagnostic and Interventional Radiology, Tuebingen University Hospital, Hoppe-Seyler-Straße 3, 72076, Tuebingen, Germany
| | - Jan Michael Brendel
- Department of Diagnostic and Interventional Radiology, Tuebingen University Hospital, Hoppe-Seyler-Straße 3, 72076, Tuebingen, Germany
| | - Konstantin Nikolaou
- Department of Diagnostic and Interventional Radiology, Tuebingen University Hospital, Hoppe-Seyler-Straße 3, 72076, Tuebingen, Germany
| | - Saif Afat
- Department of Diagnostic and Interventional Radiology, Tuebingen University Hospital, Hoppe-Seyler-Straße 3, 72076, Tuebingen, Germany
| |
Collapse
|
2
|
Patterson F, Kunar MA. The message matters: changes to binary Computer Aided Detection recommendations affect cancer detection in low prevalence search. Cogn Res Princ Implic 2024; 9:59. [PMID: 39218972 PMCID: PMC11366737 DOI: 10.1186/s41235-024-00576-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2023] [Accepted: 07/09/2024] [Indexed: 09/04/2024] Open
Abstract
Computer Aided Detection (CAD) has been used to help readers find cancers in mammograms. Although these automated systems have been shown to help cancer detection when accurate, the presence of CAD also leads to an over-reliance effect where miss errors and false alarms increase when the CAD system fails. Previous research investigated CAD systems which overlayed salient exogenous cues onto the image to highlight suspicious areas. These salient cues capture attention which may exacerbate the over-reliance effect. Furthermore, overlaying CAD cues directly on the mammogram occludes sections of breast tissue which may disrupt global statistics useful for cancer detection. In this study we investigated whether an over-reliance effect occurred with a binary CAD system, which instead of overlaying a CAD cue onto the mammogram, reported a message alongside the mammogram indicating the possible presence of a cancer. We manipulated the certainty of the message and whether it was presented only to indicate the presence of a cancer, or whether a message was displayed on every mammogram to state whether a cancer was present or absent. The results showed that although an over-reliance effect still occurred with binary CAD systems miss errors were reduced when the CAD message was more definitive and only presented to alert readers of a possible cancer.
Collapse
Affiliation(s)
| | - Melina A Kunar
- Department of Psychology, The University of Warwick, Coventry, CV4 7AL, UK
| |
Collapse
|
3
|
van der Zander QEW, Roumans R, Kusters CHJ, Dehghani N, Masclee AAM, de With PHN, van der Sommen F, Snijders CCP, Schoon EJ. Appropriate trust in artificial intelligence for the optical diagnosis of colorectal polyps: The role of human/artificial intelligence interaction. Gastrointest Endosc 2024:S0016-5107(24)03324-8. [PMID: 38942330 DOI: 10.1016/j.gie.2024.06.029] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/12/2023] [Revised: 03/26/2024] [Accepted: 06/19/2024] [Indexed: 06/30/2024]
Abstract
BACKGROUND AND AIMS Computer-aided diagnosis (CADx) for the optical diagnosis of colorectal polyps is thoroughly investigated. However, studies on human-artificial intelligence interaction are lacking. Our aim was to investigate endoscopists' trust in CADx by evaluating whether communicating a calibrated algorithm confidence score improved trust. METHODS Endoscopists optically diagnosed 60 colorectal polyps. Initially, endoscopists diagnosed the polyps without CADx assistance (initial diagnosis). Immediately afterward, the same polyp was again shown with a CADx prediction: either only a prediction (benign or premalignant) or a prediction accompanied by a calibrated confidence score (0-100). A confidence score of 0 indicated a benign prediction, 100 a (pre)malignant prediction. In half of the polyps, CADx was mandatory, and for the other half, CADx was optional. After reviewing the CADx prediction, endoscopists made a final diagnosis. Histopathology was used as the gold standard. Endoscopists' trust in CADx was measured as CADx prediction utilization: the willingness to follow CADx predictions when the endoscopists initially disagreed with the CADx prediction. RESULTS Twenty-three endoscopists participated. Presenting CADx predictions increased the endoscopists' diagnostic accuracy (69.3% initial vs 76.6% final diagnosis, P < .001). The CADx prediction was used in 36.5% (n = 183 of 501) disagreements. Adding a confidence score led to lower CADx prediction utilization, except when the confidence score surpassed 60. Mandatory CADx decreased CADx prediction utilization compared to optional CADx. Appropriate trust-using correct or disregarding incorrect CADx predictions-was 48.7% (n = 244 of 501). CONCLUSIONS Appropriate trust was common, and CADx prediction utilization was highest for the optional CADx without confidence scores. These results express the importance of a better understanding of human-artificial intelligence interaction.
Collapse
Affiliation(s)
- Quirine E W van der Zander
- Department of Gastroenterology and Hepatology, Maastricht University Medical Center, Maastricht, The Netherlands; GROW, School for Oncology and Reproduction, Maastricht University, Maastricht, The Netherlands.
| | - Rachel Roumans
- Human-Technology Interaction, Eindhoven University of Technology, Eindhoven, The Netherlands
| | - Carolus H J Kusters
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven The Netherlands
| | - Nikoo Dehghani
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven The Netherlands
| | - Ad A M Masclee
- Department of Gastroenterology and Hepatology, Maastricht University Medical Center, Maastricht, The Netherlands
| | - Peter H N de With
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven The Netherlands
| | - Fons van der Sommen
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven The Netherlands
| | - Chris C P Snijders
- Human-Technology Interaction, Eindhoven University of Technology, Eindhoven, The Netherlands
| | - Erik J Schoon
- Department of Gastroenterology and Hepatology, Maastricht University Medical Center, Maastricht, The Netherlands; Division of Gastroenterology and Hepatology, Catharina Hospital Eindhoven, Eindhoven, The Netherlands
| |
Collapse
|
4
|
Ewals LJS, Heesterbeek LJJ, Yu B, van der Wulp K, Mavroeidis D, Funk M, Snijders CCP, Jacobs I, Nederend J, Pluyter JR. The Impact of Expectation Management and Model Transparency on Radiologists' Trust and Utilization of AI Recommendations for Lung Nodule Assessment on Computed Tomography: Simulated Use Study. JMIR AI 2024; 3:e52211. [PMID: 38875574 PMCID: PMC11041414 DOI: 10.2196/52211] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/27/2023] [Revised: 11/14/2023] [Accepted: 02/03/2024] [Indexed: 06/16/2024]
Abstract
BACKGROUND Many promising artificial intelligence (AI) and computer-aided detection and diagnosis systems have been developed, but few have been successfully integrated into clinical practice. This is partially owing to a lack of user-centered design of AI-based computer-aided detection or diagnosis (AI-CAD) systems. OBJECTIVE We aimed to assess the impact of different onboarding tutorials and levels of AI model explainability on radiologists' trust in AI and the use of AI recommendations in lung nodule assessment on computed tomography (CT) scans. METHODS In total, 20 radiologists from 7 Dutch medical centers performed lung nodule assessment on CT scans under different conditions in a simulated use study as part of a 2×2 repeated-measures quasi-experimental design. Two types of AI onboarding tutorials (reflective vs informative) and 2 levels of AI output (black box vs explainable) were designed. The radiologists first received an onboarding tutorial that was either informative or reflective. Subsequently, each radiologist assessed 7 CT scans, first without AI recommendations. AI recommendations were shown to the radiologist, and they could adjust their initial assessment. Half of the participants received the recommendations via black box AI output and half received explainable AI output. Mental model and psychological trust were measured before onboarding, after onboarding, and after assessing the 7 CT scans. We recorded whether radiologists changed their assessment on found nodules, malignancy prediction, and follow-up advice for each CT assessment. In addition, we analyzed whether radiologists' trust in their assessments had changed based on the AI recommendations. RESULTS Both variations of onboarding tutorials resulted in a significantly improved mental model of the AI-CAD system (informative P=.01 and reflective P=.01). After using AI-CAD, psychological trust significantly decreased for the group with explainable AI output (P=.02). On the basis of the AI recommendations, radiologists changed the number of reported nodules in 27 of 140 assessments, malignancy prediction in 32 of 140 assessments, and follow-up advice in 12 of 140 assessments. The changes were mostly an increased number of reported nodules, a higher estimated probability of malignancy, and earlier follow-up. The radiologists' confidence in their found nodules changed in 82 of 140 assessments, in their estimated probability of malignancy in 50 of 140 assessments, and in their follow-up advice in 28 of 140 assessments. These changes were predominantly increases in confidence. The number of changed assessments and radiologists' confidence did not significantly differ between the groups that received different onboarding tutorials and AI outputs. CONCLUSIONS Onboarding tutorials help radiologists gain a better understanding of AI-CAD and facilitate the formation of a correct mental model. If AI explanations do not consistently substantiate the probability of malignancy across patient cases, radiologists' trust in the AI-CAD system can be impaired. Radiologists' confidence in their assessments was improved by using the AI recommendations.
Collapse
Affiliation(s)
- Lotte J S Ewals
- Catharina Cancer Institute, Catharina Hospital Eindhoven, Eindhoven, Netherlands
| | | | - Bin Yu
- Research Center for Marketing and Supply Chain Management, Nyenrode Business University, Breukelen, Netherlands
| | - Kasper van der Wulp
- Catharina Cancer Institute, Catharina Hospital Eindhoven, Eindhoven, Netherlands
| | | | - Mathias Funk
- Department of Industrial Design, Eindhoven University of Technology, Eindhoven, Netherlands
| | - Chris C P Snijders
- Department of Human Technology Interaction, Eindhoven University of Technology, Eindhoven, Netherlands
| | - Igor Jacobs
- Department of Hospital Services and Informatics, Philips Research, Eindhoven, Netherlands
| | - Joost Nederend
- Catharina Cancer Institute, Catharina Hospital Eindhoven, Eindhoven, Netherlands
| | - Jon R Pluyter
- Department of Experience Design, Royal Philips, Eindhoven, Netherlands
| |
Collapse
|
5
|
Lombi L, Rossero E. How artificial intelligence is reshaping the autonomy and boundary work of radiologists. A qualitative study. SOCIOLOGY OF HEALTH & ILLNESS 2024; 46:200-218. [PMID: 37573551 DOI: 10.1111/1467-9566.13702] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/26/2023] [Accepted: 07/19/2023] [Indexed: 08/15/2023]
Abstract
The application of artificial intelligence (AI) in medical practice is spreading, especially in technologically dense fields such as radiology, which could consequently undergo profound transformations in the near future. This article aims to qualitatively explore the potential influence of AI technologies on the professional identity of radiologists. Drawing on 12 in-depth interviews with a subgroup of radiologists who participated in a larger study, this article investigated (1) whether radiologists perceived AI as a threat to their decision-making autonomy; and (2) how radiologists perceived the future of their profession compared to other health-care professions. The findings revealed that while AI did not generally affect radiologists' decision-making autonomy, it threatened their professional and epistemic authority. Two discursive strategies were identified to explain these findings. The first strategy emphasised radiologists' specific expertise and knowledge that extends beyond interpreting images, a task performed with high accuracy by AI machines. The second strategy underscored the fostering of radiologists' professional prestige through developing expertise in using AI technologies, a skill that would distinguish them from other clinicians who did not pose this knowledge. This study identifies AI machines as status objects and useful tools in performing boundary work in and around the radiological profession.
Collapse
Affiliation(s)
- Linda Lombi
- Department of Sociology, Università Cattolica del Sacro Cuore, Milan, Italy
| | - Eleonora Rossero
- Fundamental Rights Laboratory, Collegio Carlo Alberto, Turin, Italy
| |
Collapse
|
6
|
Rezazade Mehrizi MH, Mol F, Peter M, Ranschaert E, Dos Santos DP, Shahidi R, Fatehi M, Dratsch T. The impact of AI suggestions on radiologists' decisions: a pilot study of explainability and attitudinal priming interventions in mammography examination. Sci Rep 2023; 13:9230. [PMID: 37286665 DOI: 10.1038/s41598-023-36435-3] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2022] [Accepted: 06/03/2023] [Indexed: 06/09/2023] Open
Abstract
Various studies have shown that medical professionals are prone to follow the incorrect suggestions offered by algorithms, especially when they have limited inputs to interrogate and interpret such suggestions and when they have an attitude of relying on them. We examine the effect of correct and incorrect algorithmic suggestions on the diagnosis performance of radiologists when (1) they have no, partial, and extensive informational inputs for explaining the suggestions (study 1) and (2) they are primed to hold a positive, negative, ambivalent, or neutral attitude towards AI (study 2). Our analysis of 2760 decisions made by 92 radiologists conducting 15 mammography examinations shows that radiologists' diagnoses follow both incorrect and correct suggestions, despite variations in the explainability inputs and attitudinal priming interventions. We identify and explain various pathways through which radiologists navigate through the decision process and arrive at correct or incorrect decisions. Overall, the findings of both studies show the limited effect of using explainability inputs and attitudinal priming for overcoming the influence of (incorrect) algorithmic suggestions.
Collapse
Affiliation(s)
| | - Ferdinand Mol
- Vrije Universiteit Amsterdam, Amsterdam, The Netherlands
| | - Marcel Peter
- Vrije Universiteit Amsterdam, Amsterdam, The Netherlands
| | | | - Daniel Pinto Dos Santos
- Institute of Diagnostic and Interventional Radiology, Faculty of Medicine and University Hospital Cologne, University of Cologne, Cologne, Germany
| | - Ramin Shahidi
- Bushehr University of Medical Sciences, Bushehr, Iran
| | | | - Thomas Dratsch
- Institute of Diagnostic and Interventional Radiology, Faculty of Medicine and University Hospital Cologne, University of Cologne, Cologne, Germany
| |
Collapse
|
7
|
Boopathiraja S, Kalavathi P, Deoghare S, Prasath VBS. Near Lossless Compression for 3D Radiological Images Using Optimal Multilinear Singular Value Decomposition (3D-VOI-OMLSVD). J Digit Imaging 2023; 36:259-275. [PMID: 36038701 PMCID: PMC9422948 DOI: 10.1007/s10278-022-00687-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2022] [Revised: 07/14/2022] [Accepted: 07/18/2022] [Indexed: 11/03/2022] Open
Abstract
Storage and transmission of high-compression 3D radiological images that create high-quality reconstruction upon decompression are critical necessities for effective and efficient teleradiology. To cater to this need, we propose a near lossless 3D image volume compression method based on optimal multilinear singular value decomposition called "3D-VOI-OMLSVD." The proposed strategy first eliminates any blank 2D image slices from the 3D image volume and uses the selective bounding volume (SBV) to identify and extract the volume of Interest (VOI). Following this, the VOI is decomposed with an optimal multilinear singular value decomposition (OMLSVD) to obtain the corresponding core tensor, factor matrices, and singular values that are compressed with adaptive binary range coder (ABRC), integrated as an entropy encoder. The compressed file can be transferred or transmitted and then decompressed in order to reconstruct the original image. The resultant decompressed VOI is acquired by reversing the above process and then fusing it with the background, using the bound volume coordinates associated with the compressed 3D image. The proposed method performance was tested on a variety of 3D radiological images with different imaging modalities and dimensions using quantitative evaluation metrics such as the compression rate (CR), bit rate (BR), peak signal to noise ratio (PSNR), and structural similarity index (SSIM). Furthermore, we also investigate the impact of VOI extraction on the model performance, before comparing it with two popular compression methods, namely JPEG and JPEG2000. Our proposed method, 3D-VOI-OMLSVD, displayed a high CR value, with a maximum of 37.31, and a low BR, with the lowest reported to be 0.21. The SSIM score was consistently high, with an average performance of 0.9868, while using < 1 second for decoding the image. We observe that with VOI extraction, the compression rate increases manifold, and bit rate drops significantly, and thus reduces the encoding and decoding time to a great extent. Compared to JPEG and JPEG2000, our method consistently performs better in terms of higher CR and lower BR. The results indicate that the proposed compression methodology performs consistently to create high-quality image compressions, and overall gives a better outcome when compared against two state-of-the-art and widely used methods, JPEG and JPEG2000.
Collapse
Affiliation(s)
- S. Boopathiraja
- Department of Computer Science and Applications, The Gandhigram Rural Institute (Deemed to Be University), Gandhigram, 624 302 Tamil Nadu India
| | - P. Kalavathi
- Department of Computer Science and Applications, The Gandhigram Rural Institute (Deemed to Be University), Gandhigram, 624 302 Tamil Nadu India
| | - S. Deoghare
- Division of Biomedical Informatics, Cincinnati Children’s Hospital Medical Center, Cincinnati, OH 45229 USA
- Department of Biomedical Informatics, College of Medicine, University of Cincinnati, Cincinnati, OH 45267 USA
| | - V. B. Surya Prasath
- Division of Biomedical Informatics, Cincinnati Children’s Hospital Medical Center, Cincinnati, OH 45229 USA
- Department of Biomedical Informatics, College of Medicine, University of Cincinnati, Cincinnati, OH 45267 USA
- Department of Pediatrics, College of Medicine, University of Cincinnati, Cincinnati, OH 45257 USA
- Department of Electrical Engineering and Computer Science, College of Engineering and Applied Science, University of Cincinnati, Cincinnati, OH 45221 USA
| |
Collapse
|
8
|
Reverberi C, Rigon T, Solari A, Hassan C, Cherubini P, Cherubini A. Experimental evidence of effective human-AI collaboration in medical decision-making. Sci Rep 2022; 12:14952. [PMID: 36056152 PMCID: PMC9440124 DOI: 10.1038/s41598-022-18751-2] [Citation(s) in RCA: 19] [Impact Index Per Article: 9.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2022] [Accepted: 08/18/2022] [Indexed: 11/25/2022] Open
Abstract
Artificial Intelligence (AI) systems are precious support for decision-making, with many applications also in the medical domain. The interaction between MDs and AI enjoys a renewed interest following the increased possibilities of deep learning devices. However, we still have limited evidence-based knowledge of the context, design, and psychological mechanisms that craft an optimal human-AI collaboration. In this multicentric study, 21 endoscopists reviewed 504 videos of lesions prospectively acquired from real colonoscopies. They were asked to provide an optical diagnosis with and without the assistance of an AI support system. Endoscopists were influenced by AI ([Formula: see text]), but not erratically: they followed the AI advice more when it was correct ([Formula: see text]) than incorrect ([Formula: see text]). Endoscopists achieved this outcome through a weighted integration of their and the AI opinions, considering the case-by-case estimations of the two reliabilities. This Bayesian-like rational behavior allowed the human-AI hybrid team to outperform both agents taken alone. We discuss the features of the human-AI interaction that determined this favorable outcome.
Collapse
Affiliation(s)
- Carlo Reverberi
- Department of Psychology, University of Milano-Bicocca, 20126, Milan, Italy.
- Milan Center for Neuroscience, University of Milano-Bicocca, 20126, Milan, Italy.
| | - Tommaso Rigon
- Department of Economics, Management and Statistics, University of Milano-Bicocca, 20126, Milan, Italy
| | - Aldo Solari
- Milan Center for Neuroscience, University of Milano-Bicocca, 20126, Milan, Italy
- Department of Economics, Management and Statistics, University of Milano-Bicocca, 20126, Milan, Italy
| | - Cesare Hassan
- Department of Biomedical Sciences, Humanitas University, 20072, Pieve Emanuele, Italy
- Endoscopy Unit, Humanitas Clinical and Research Center IRCCS, Rozzano, Italy
| | - Paolo Cherubini
- Department of Psychology, University of Milano-Bicocca, 20126, Milan, Italy
- Milan Center for Neuroscience, University of Milano-Bicocca, 20126, Milan, Italy
- Department of Neural and Behavioral Sciences, University of Pavia, Pavia, Italy
| | - Andrea Cherubini
- Milan Center for Neuroscience, University of Milano-Bicocca, 20126, Milan, Italy.
- Artificial Intelligence Group, Cosmo AI/Linkverse, Lainate, 20045, Milan, Italy.
| |
Collapse
|
9
|
Integrating patient symptoms, clinical readings, and radiologist feedback with computer-aided diagnosis system for detection of infectious pulmonary disease: a feasibility study. Med Biol Eng Comput 2022; 60:2549-2565. [DOI: 10.1007/s11517-022-02611-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2021] [Accepted: 06/07/2022] [Indexed: 10/17/2022]
|
10
|
Deep Learning–Based Brain Computed Tomography Image Classification with Hyperparameter Optimization through Transfer Learning for Stroke. Diagnostics (Basel) 2022; 12:diagnostics12040807. [PMID: 35453855 PMCID: PMC9026481 DOI: 10.3390/diagnostics12040807] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2022] [Revised: 03/21/2022] [Accepted: 03/24/2022] [Indexed: 02/04/2023] Open
Abstract
Brain computed tomography (CT) is commonly used for evaluating the cerebral condition, but immediately and accurately interpreting emergent brain CT images is tedious, even for skilled neuroradiologists. Deep learning networks are commonly employed for medical image analysis because they enable efficient computer-aided diagnosis. This study proposed the use of convolutional neural network (CNN)-based deep learning models for efficient classification of strokes based on unenhanced brain CT image findings into normal, hemorrhage, infarction, and other categories. The included CNN models were CNN-2, VGG-16, and ResNet-50, all of which were pretrained through transfer learning with various data sizes, mini-batch sizes, and optimizers. Their performance in classifying unenhanced brain CT images was tested thereafter. This performance was then compared with the outcomes in other studies on deep learning–based hemorrhagic or ischemic stroke diagnoses. The results revealed that among our CNN-2, VGG-16, and ResNet-50 analyzed by considering several hyperparameters and environments, the CNN-2 and ResNet-50 outperformed the VGG-16, with an accuracy of 0.9872; however, ResNet-50 required a longer time to present the outcome than did the other networks. Moreover, our models performed much better than those reported previously. In conclusion, after appropriate hyperparameter optimization, our deep learning–based models can be applied to clinical scenarios where neurologist or radiologist may need to verify whether their patients have a hemorrhage stroke, an infarction, and any other symptom.
Collapse
|
11
|
Basso MN, Barua M, John R, Khademi A. Explainable Biomarkers for Automated Glomerular and Patient-Level Disease Classification. KIDNEY360 2021; 3:534-545. [PMID: 35582169 PMCID: PMC9034815 DOI: 10.34067/kid.0005102021] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/05/2021] [Accepted: 12/08/2021] [Indexed: 01/12/2023]
Abstract
Pathologists use multiple microscopy modalities to assess renal biopsy specimens. Besides usual diagnostic features, some changes are too subtle to be properly defined. Computational approaches have the potential to systematically quantitate subvisual clues, provide pathogenetic insight, and link to clinical outcomes. To this end, a proof-of-principle study is presented demonstrating that explainable biomarkers through machine learning can distinguish between glomerular disorders at the light-microscopy level. The proposed system used image analysis techniques and extracted 233 explainable biomarkers related to color, morphology, and microstructural texture. Traditional machine learning was then used to classify minimal change disease (MCD), membranous nephropathy (MN), and thin basement membrane nephropathy (TBMN) diseases on a glomerular and patient-level basis. The final model combined the Gini feature importance set and linear discriminant analysis classifier. Six morphologic (nuclei-to-glomerular tuft area, nuclei-to-glomerular area, glomerular tuft thickness greater than ten, glomerular tuft thickness greater than three, total glomerular tuft thickness, and glomerular circularity) and four microstructural texture features (luminal contrast using wavelets, nuclei energy using wavelets, nuclei variance using color vector LBP, and glomerular correlation using GLCM) were, together, the best performing biomarkers. Accuracies of 77% and 87% were obtained for glomerular and patient-level classification, respectively. Computational methods, using explainable glomerular biomarkers, have diagnostic value and are compatible with our existing knowledge of disease pathogenesis. Furthermore, this algorithm can be applied to clinical datasets for novel prognostic and mechanistic biomarker discovery.
Collapse
Affiliation(s)
- Matthew Nicholas Basso
- Image Analysis in Medicine Lab (IAMLAB), Department of Electrical, Computer, and Biomedical Engineering, Ryerson University, Toronto, Canada
| | - Moumita Barua
- Division of Nephrology, University Health Network, Toronto, Canada,Toronto General Hospital Research Institute, Toronto General Hospital, Toronto, Canada,Department of Medicine, University of Toronto, Toronto, Canada,Institute of Medical Sciences, University of Toronto, Toronto, Canada
| | - Rohan John
- Department of Pathology, University Health Network, Toronto, Canada
| | - April Khademi
- Image Analysis in Medicine Lab (IAMLAB), Department of Electrical, Computer, and Biomedical Engineering, Ryerson University, Toronto, Canada,Keenan Research Center for Biomedical Science, St. Michael’s Hospital, Unity Health Network, Toronto, Canada,Institute for Biomedical Engineering, Science, and Technology (iBEST), a partnership between St. Michael’s Hospital and Ryerson University, Toronto, Canada
| |
Collapse
|
12
|
Zhou SK, Greenspan H, Davatzikos C, Duncan JS, van Ginneken B, Madabhushi A, Prince JL, Rueckert D, Summers RM. A review of deep learning in medical imaging: Imaging traits, technology trends, case studies with progress highlights, and future promises. PROCEEDINGS OF THE IEEE. INSTITUTE OF ELECTRICAL AND ELECTRONICS ENGINEERS 2021; 109:820-838. [PMID: 37786449 PMCID: PMC10544772 DOI: 10.1109/jproc.2021.3054390] [Citation(s) in RCA: 229] [Impact Index Per Article: 76.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/04/2023]
Abstract
Since its renaissance, deep learning has been widely used in various medical imaging tasks and has achieved remarkable success in many medical imaging applications, thereby propelling us into the so-called artificial intelligence (AI) era. It is known that the success of AI is mostly attributed to the availability of big data with annotations for a single task and the advances in high performance computing. However, medical imaging presents unique challenges that confront deep learning approaches. In this survey paper, we first present traits of medical imaging, highlight both clinical needs and technical challenges in medical imaging, and describe how emerging trends in deep learning are addressing these issues. We cover the topics of network architecture, sparse and noisy labels, federating learning, interpretability, uncertainty quantification, etc. Then, we present several case studies that are commonly found in clinical practice, including digital pathology and chest, brain, cardiovascular, and abdominal imaging. Rather than presenting an exhaustive literature survey, we instead describe some prominent research highlights related to these case study applications. We conclude with a discussion and presentation of promising future directions.
Collapse
Affiliation(s)
- S Kevin Zhou
- School of Biomedical Engineering, University of Science and Technology of China and Institute of Computing Technology, Chinese Academy of Sciences
| | - Hayit Greenspan
- Biomedical Engineering Department, Tel-Aviv University, Israel
| | - Christos Davatzikos
- Radiology Department and Electrical and Systems Engineering Department, University of Pennsylvania, USA
| | - James S Duncan
- Departments of Biomedical Engineering and Radiology & Biomedical Imaging, Yale University
| | | | - Anant Madabhushi
- Department of Biomedical Engineering, Case Western Reserve University and Louis Stokes Cleveland Veterans Administration Medical Center, USA
| | - Jerry L Prince
- Electrical and Computer Engineering Department, Johns Hopkins University, USA
| | - Daniel Rueckert
- Klinikum rechts der Isar, TU Munich, Germany and Department of Computing, Imperial College, UK
| | | |
Collapse
|
13
|
Cicalese PA, Mobiny A, Shahmoradi Z, Yi X, Mohan C, Van Nguyen H. Kidney Level Lupus Nephritis Classification Using Uncertainty Guided Bayesian Convolutional Neural Networks. IEEE J Biomed Health Inform 2021; 25:315-324. [PMID: 33206612 DOI: 10.1109/jbhi.2020.3039162] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
The kidney biopsy based diagnosis of Lupus Nephritis (LN) is characterized by low inter-observer agreement, with misdiagnosis being associated with increased patient morbidity and mortality. Although various Computer Aided Diagnosis (CAD) systems have been developed for other nephrohistopathological applications, little has been done to accurately classify kidneys based on their kidney level Lupus Glomerulonephritis (LGN) scores. The successful implementation of CAD systems has also been hindered by the diagnosing physician's perceived classifier strengths and weaknesses, which has been shown to have a negative effect on patient outcomes. We propose an Uncertainty-Guided Bayesian Classification (UGBC) scheme that is designed to accurately classify control, class I/II, and class III/IV LGN (3 class) at both the glomerular-level classification task (26,634 segmented glomerulus images) and the kidney-level classification task (87 MRL/lpr mouse kidney sections). Data annotation was performed using a high throughput, bulk labeling scheme that is designed to take advantage of Deep Neural Network's (or DNNs) resistance to label noise. Our augmented UGBC scheme achieved a 94.5% weighted glomerular-level accuracy while achieving a weighted kidney-level accuracy of 96.6%, improving upon the standard Convolutional Neural Network (CNN) architecture by 11.8% and 3.5% respectively.
Collapse
|
14
|
Gryska E, Schneiderman J, Björkman-Burtscher I, Heckemann RA. Automatic brain lesion segmentation on standard magnetic resonance images: a scoping review. BMJ Open 2021; 11:e042660. [PMID: 33514580 PMCID: PMC7849889 DOI: 10.1136/bmjopen-2020-042660] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/11/2020] [Revised: 01/09/2021] [Accepted: 01/12/2021] [Indexed: 12/11/2022] Open
Abstract
OBJECTIVES Medical image analysis practices face challenges that can potentially be addressed with algorithm-based segmentation tools. In this study, we map the field of automatic MR brain lesion segmentation to understand the clinical applicability of prevalent methods and study designs, as well as challenges and limitations in the field. DESIGN Scoping review. SETTING Three databases (PubMed, IEEE Xplore and Scopus) were searched with tailored queries. Studies were included based on predefined criteria. Emerging themes during consecutive title, abstract, methods and whole-text screening were identified. The full-text analysis focused on materials, preprocessing, performance evaluation and comparison. RESULTS Out of 2990 unique articles identified through the search, 441 articles met the eligibility criteria, with an estimated growth rate of 10% per year. We present a general overview and trends in the field with regard to publication sources, segmentation principles used and types of lesions. Algorithms are predominantly evaluated by measuring the agreement of segmentation results with a trusted reference. Few articles describe measures of clinical validity. CONCLUSIONS The observed reporting practices leave room for improvement with a view to studying replication, method comparison and clinical applicability. To promote this improvement, we propose a list of recommendations for future studies in the field.
Collapse
Affiliation(s)
- Emilia Gryska
- Medical Radiation Sciences, Goteborgs universitet Institutionen for kliniska vetenskaper, Goteborg, Sweden
| | - Justin Schneiderman
- Sektionen för klinisk neurovetenskap, Goteborgs Universitet Institutionen for Neurovetenskap och fysiologi, Goteborg, Sweden
| | | | - Rolf A Heckemann
- Medical Radiation Sciences, Goteborgs universitet Institutionen for kliniska vetenskaper, Goteborg, Sweden
| |
Collapse
|
15
|
Jackson BR, Ye Y, Crawford JM, Becich MJ, Roy S, Botkin JR, de Baca ME, Pantanowitz L. The Ethics of Artificial Intelligence in Pathology and Laboratory Medicine: Principles and Practice. Acad Pathol 2021; 8:2374289521990784. [PMID: 33644301 PMCID: PMC7894680 DOI: 10.1177/2374289521990784] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2020] [Revised: 11/24/2020] [Accepted: 12/28/2020] [Indexed: 12/24/2022] Open
Abstract
Growing numbers of artificial intelligence applications are being developed and applied to pathology and laboratory medicine. These technologies introduce risks and benefits that must be assessed and managed through the lens of ethics. This article describes how long-standing principles of medical and scientific ethics can be applied to artificial intelligence using examples from pathology and laboratory medicine.
Collapse
Affiliation(s)
- Brian R. Jackson
- Department of Pathology, University of Utah School of Medicine, Salt Lake City, UT, USA
- ARUP Laboratories, Salt Lake City, UT, USA
| | - Ye Ye
- Department of Biomedical Informatics, University of Pittsburgh School of Medicine, Pittsburgh, PA, USA
| | - James M. Crawford
- Department of Pathology and Laboratory Medicine, Donald and Barbara Zucker School of Medicine at Hofstra/Northwell, Hempstead, NY, USA
| | - Michael J. Becich
- Department of Biomedical Informatics, University of Pittsburgh School of Medicine, Pittsburgh, PA, USA
| | - Somak Roy
- Division of Pathology, Cincinnati Children’s Hospital Medical Center, Cincinnati, OH, USA
| | - Jeffrey R. Botkin
- Department of Pediatrics, University of Utah School of Medicine, Salt Lake City, UT, USA
| | | | | |
Collapse
|
16
|
Kondylakis H, Axenie C, Kiran Bastola D, Katehakis DG, Kouroubali A, Kurz D, Larburu N, Macía I, Maguire R, Maramis C, Marias K, Morrow P, Muro N, Núñez-Benjumea FJ, Rampun A, Rivera-Romero O, Scotney B, Signorelli G, Wang H, Tsiknakis M, Zwiggelaar R. Status and Recommendations of Technological and Data-Driven Innovations in Cancer Care: Focus Group Study. J Med Internet Res 2020; 22:e22034. [PMID: 33320099 PMCID: PMC7772066 DOI: 10.2196/22034] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2020] [Revised: 10/02/2020] [Accepted: 10/26/2020] [Indexed: 12/16/2022] Open
Abstract
BACKGROUND The status of the data-driven management of cancer care as well as the challenges, opportunities, and recommendations aimed at accelerating the rate of progress in this field are topics of great interest. Two international workshops, one conducted in June 2019 in Cordoba, Spain, and one in October 2019 in Athens, Greece, were organized by four Horizon 2020 (H2020) European Union (EU)-funded projects: BOUNCE, CATCH ITN, DESIREE, and MyPal. The issues covered included patient engagement, knowledge and data-driven decision support systems, patient journey, rehabilitation, personalized diagnosis, trust, assessment of guidelines, and interoperability of information and communication technology (ICT) platforms. A series of recommendations was provided as the complex landscape of data-driven technical innovation in cancer care was portrayed. OBJECTIVE This study aims to provide information on the current state of the art of technology and data-driven innovations for the management of cancer care through the work of four EU H2020-funded projects. METHODS Two international workshops on ICT in the management of cancer care were held, and several topics were identified through discussion among the participants. A focus group was formulated after the second workshop, in which the status of technological and data-driven cancer management as well as the challenges, opportunities, and recommendations in this area were collected and analyzed. RESULTS Technical and data-driven innovations provide promising tools for the management of cancer care. However, several challenges must be successfully addressed, such as patient engagement, interoperability of ICT-based systems, knowledge management, and trust. This paper analyzes these challenges, which can be opportunities for further research and practical implementation and can provide practical recommendations for future work. CONCLUSIONS Technology and data-driven innovations are becoming an integral part of cancer care management. In this process, specific challenges need to be addressed, such as increasing trust and engaging the whole stakeholder ecosystem, to fully benefit from these innovations.
Collapse
Affiliation(s)
| | - Cristian Axenie
- Audi Konfuzius-Institut Ingolstadt Lab, Technische Hochschule Ingolstadt, Ingolstadt, Germany
| | - Dhundy Kiran Bastola
- School of Interdisciplinary Informatics, University of Nebraska, Omaha, NE, United States
| | | | | | - Daria Kurz
- Interdisziplinäres Brustzentrum, Helios Klinikum München West, Munich, Germany
| | - Nekane Larburu
- Vicomtech, Health Research Institute, San Sebastian, Spain
| | - Iván Macía
- Vicomtech, Health Research Institute, San Sebastian, Spain
| | - Roma Maguire
- University of Strathclyde, Glasgow, United Kingdom
| | - Christos Maramis
- eHealth Lab, Institute of Applied Biosciences - Centre for Research & Technology Hellas, Thessaloniki, Greece
| | | | - Philip Morrow
- School of Computing, Ulster University, Newtownabbey, United Kingdom
| | - Naiara Muro
- Vicomtech, Health Research Institute, San Sebastian, Spain
| | | | - Andrik Rampun
- Academic Unit of Radiology, Department of Infection, Immunity and Cardiovascular Disease, University of Sheffield, Sheffield, United Kingdom
| | | | - Bryan Scotney
- School of Computing, Ulster University, Newtownabbey, United Kingdom
| | | | - Hui Wang
- School of Computing and Engineering, University of West London, London, United Kingdom
| | | | - Reyer Zwiggelaar
- Department of Computer Science, Aberystwyth University, Aberystwyth, United Kingdom
| |
Collapse
|
17
|
Filice RW, Ratwani RM. The Case for User-Centered Artificial Intelligence in Radiology. Radiol Artif Intell 2020; 2:e190095. [PMID: 33937824 PMCID: PMC8082296 DOI: 10.1148/ryai.2020190095] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2019] [Revised: 12/14/2019] [Accepted: 01/06/2020] [Indexed: 06/12/2023]
Abstract
Past technology transition successes and failures have demonstrated the importance of user-centered design and the science of human factors; these approaches will be critical to the success of artificial intelligence in radiology.
Collapse
Affiliation(s)
- Ross W. Filice
- From MedStar Health, MedStar Georgetown University Hospital, 3800 Reservoir Rd, NW CG201, Washington DC, 20007 (R.W.F.); and MedStar Health, National Center for Human Factors in Healthcare, Washington, DC (R.M.R.)
| | - Raj M. Ratwani
- From MedStar Health, MedStar Georgetown University Hospital, 3800 Reservoir Rd, NW CG201, Washington DC, 20007 (R.W.F.); and MedStar Health, National Center for Human Factors in Healthcare, Washington, DC (R.M.R.)
| |
Collapse
|
18
|
Mascarenhas LR, Ribeiro Júnior ADS, Ramos RP. Automatic segmentation of brain tumors in magnetic resonance imaging. EINSTEIN-SAO PAULO 2020; 18:eAO4948. [PMID: 32159604 PMCID: PMC7053828 DOI: 10.31744/einstein_journal/2020ao4948] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2019] [Accepted: 09/02/2019] [Indexed: 11/21/2022] Open
Abstract
Objective To develop a computational algorithm applied to magnetic resonance imaging for automatic segmentation of brain tumors. Methods A total of 130 magnetic resonance images were used in the T1c, T2 and FSPRG T1C sequences and in the axial, sagittal and coronal planes of patients with brain cancer. The algorithms employed contrast correction, histogram normalization and binarization techniques to disconnect adjacent structures from the brain and enhance the region of interest. Automatic segmentation was performed through detection by coordinates and arithmetic mean of the area. Morphological operators were used to eliminate undesirable elements and reconstruct the shape and texture of the tumor. The results were compared with manual segmentations by two radiologists to determine the efficacy of the algorithms implemented. Results The correlated correspondence between the segmentation obtained and the gold standard was 89.23%. Conclusion It is possible to locate and define the tumor region automatically with no the need for user interaction, based on two innovative methods to detect brain extreme sites and exclude non-tumor tissues on magnetic resonance images.
Collapse
|
19
|
Du-Crow E, Astley SM, Hulleman J. Is there a safety-net effect with computer-aided detection? J Med Imaging (Bellingham) 2020; 7:022405. [PMID: 31903408 PMCID: PMC6931663 DOI: 10.1117/1.jmi.7.2.022405] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2019] [Accepted: 12/05/2019] [Indexed: 11/14/2022] Open
Abstract
Computer-aided detection (CAD) systems are used to aid readers interpreting screening mammograms. An expert reader searches the image initially unaided and then once again with the aid of CAD, which prompts automatically detected suspicious regions. This could lead to a "safety-net" effect, where the initial unaided search of the image is adversely affected by the fact that it is preliminary to an additional search with CAD and may, therefore, be less thorough. To investigate the existence of such an effect, we created a visual search experiment for nonexpert observers mirroring breast screening with CAD. Each observer searched 100 images for microcalcification clusters within synthetic images in both prompted (CAD) and unprompted (no-CAD) conditions. Fifty-two participants were recruited for the study, 48 of whom had their eye movements tracked in real-time; the other 4 participants could not be accurately calibrated, so only behavioral data were collected. In the CAD condition, before prompts were displayed, image coverage was significantly lower than coverage in the no-CAD condition (t 47 = 5.29 , p < 0.0001 ). Observer sensitivity was significantly greater for targets marked by CAD than the same targets in the no-CAD condition (t 51 = 6.56 , p < 0.001 ). For targets not marked by CAD, there was no significant difference in observer sensitivity in the CAD condition compared with the same targets in the no-CAD condition (t 51 = 0.54 , p = 0.59 ). These results suggest that the initial search may be influenced by the subsequent availability of CAD; if so, cross-sectional CAD efficacy studies should account for the effect when estimating benefit.
Collapse
Affiliation(s)
- Ethan Du-Crow
- Division of Informatics, Imaging and Data Sciences, University of Manchester, Faculty of Biology, Medicine and Health, Manchester, United Kingdom
| | - Susan M Astley
- Division of Informatics, Imaging and Data Sciences, University of Manchester, Faculty of Biology, Medicine and Health, Manchester, United Kingdom
| | - Johan Hulleman
- Division of Neuroscience and Experimental Psychology, University of Manchester, Faculty of Biology, Medicine and Health, Manchester, United Kingdom
| |
Collapse
|
20
|
Chang L, Fu C, Wu Z, Liu W, Yang S. Data-Driven Analysis of Radiologists' Behavior for Diagnosing Thyroid Nodules. IEEE J Biomed Health Inform 2020; 24:3111-3123. [PMID: 32012031 DOI: 10.1109/jbhi.2020.2969322] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Thyroid nodule has been a common and serious threaten to human health. With the identification and diagnosis of thyroid nodules in the general population, large volumes of examination reports in clinical practice have been accumulated. They provide data basics of analyzing radiologists' behavior of diagnosing thyroid nodules. To conduct data-driven analysis of radiologists' behavior, an experimental framework is designed based on belief rule base, which is essentially a white box for knowledge representation and uncertain reasoning. Under the framework, with 2744 examination reports of thyroid nodules in the period from January 2012 to February 2019 that have been collected from a tertiary hospital located in Hefei, Anhui, China, experimental results are obtained from conducting missing validation, self-validation, and mutual validation. Three principles are then concluded from the results and corresponding analysis. The first is that missing features on some criteria are considered as benign ones by default, the second is that there is generally inconsistency between the recorded features on criteria and the overall diagnosis, and the third is that different radiologists have different diagnostic preferences. These three principles reflect three diagnostic behavioral characteristics of radiologists, namely reliability, inconsistency, and independence. Based on the three principles and radiologists' behavioral characteristics, managerial insights in a general case are concluded to make the findings in this study available in other situations.
Collapse
|
21
|
Mobiny A, Singh A, Van Nguyen H. Risk-Aware Machine Learning Classifier for Skin Lesion Diagnosis. J Clin Med 2019; 8:E1241. [PMID: 31426482 PMCID: PMC6723257 DOI: 10.3390/jcm8081241] [Citation(s) in RCA: 31] [Impact Index Per Article: 6.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2019] [Revised: 08/12/2019] [Accepted: 08/15/2019] [Indexed: 01/01/2023] Open
Abstract
Knowing when a machine learning system is not confident about its prediction is crucial in medical domains where safety is critical. Ideally, a machine learning algorithm should make a prediction only when it is highly certain about its competency, and refer the case to physicians otherwise. In this paper, we investigate how Bayesian deep learning can improve the performance of the machine-physician team in the skin lesion classification task. We used the publicly available HAM10000 dataset, which includes samples from seven common skin lesion categories: Melanoma (MEL), Melanocytic Nevi (NV), Basal Cell Carcinoma (BCC), Actinic Keratoses and Intraepithelial Carcinoma (AKIEC), Benign Keratosis (BKL), Dermatofibroma (DF), and Vascular (VASC) lesions. Our experimental results show that Bayesian deep networks can boost the diagnostic performance of the standard DenseNet-169 model from 81.35% to 83.59% without incurring additional parameters or heavy computation. More importantly, a hybrid physician-machine workflow reaches a classification accuracy of 90 % while only referring 35 % of the cases to physicians. The findings are expected to generalize to other medical diagnosis applications. We believe that the availability of risk-aware machine learning methods will enable a wider adoption of machine learning technology in clinical settings.
Collapse
Affiliation(s)
- Aryan Mobiny
- Department of Electrical and Computer Engineering, University of Houston, Houston, TX 77004, USA.
| | - Aditi Singh
- Department of Electrical and Computer Engineering, University of Houston, Houston, TX 77004, USA
| | - Hien Van Nguyen
- Department of Electrical and Computer Engineering, University of Houston, Houston, TX 77004, USA
| |
Collapse
|
22
|
Shen S, Han SX, Aberle DR, Bui AA, Hsu W. An Interpretable Deep Hierarchical Semantic Convolutional Neural Network for Lung Nodule Malignancy Classification. EXPERT SYSTEMS WITH APPLICATIONS 2019; 128:84-95. [PMID: 31296975 PMCID: PMC6623975 DOI: 10.1016/j.eswa.2019.01.048] [Citation(s) in RCA: 85] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
While deep learning methods have demonstrated performance comparable to human readers in tasks such as computer-aided diagnosis, these models are difficult to interpret, do not incorporate prior domain knowledge, and are often considered as a "black-box." The lack of model interpretability hinders them from being fully understood by end users such as radiologists. In this paper, we present a novel interpretable deep hierarchical semantic convolutional neural network (HSCNN) to predict whether a given pulmonary nodule observed on a computed tomography (CT) scan is malignant. Our network provides two levels of output: 1) low-level semantic features; and 2) a high-level prediction of nodule malignancy. The low-level outputs reflect diagnostic features often reported by radiologists and serve to explain how the model interprets the images in an expert-interpretable manner. The information from these low-level outputs, along with the representations learned by the convolutional layers, are then combined and used to infer the high-level output. This unified architecture is trained by optimizing a global loss function including both low- and high-level tasks, thereby learning all the parameters within a joint framework. Our experimental results using the Lung Image Database Consortium (LIDC) show that the proposed method not only produces interpretable lung cancer predictions but also achieves significantly better results compared to using a 3D CNN alone.
Collapse
Affiliation(s)
- Shiwen Shen
- Department of Bioengineering, University of California, Los Angeles, CA, USA
- Medical & Imaging Informatics Group, Department of Radiological Sciences, University of California, Los Angeles, CA, USA
| | - Simon X Han
- Department of Bioengineering, University of California, Los Angeles, CA, USA
- Medical & Imaging Informatics Group, Department of Radiological Sciences, University of California, Los Angeles, CA, USA
| | - Denise R Aberle
- Department of Bioengineering, University of California, Los Angeles, CA, USA
- Medical & Imaging Informatics Group, Department of Radiological Sciences, University of California, Los Angeles, CA, USA
| | - Alex A Bui
- Medical & Imaging Informatics Group, Department of Radiological Sciences, University of California, Los Angeles, CA, USA
| | - William Hsu
- Medical & Imaging Informatics Group, Department of Radiological Sciences, University of California, Los Angeles, CA, USA
| |
Collapse
|
23
|
Gaur S, Lay N, Harmon SA, Doddakashi S, Mehralivand S, Argun B, Barrett T, Bednarova S, Girometti R, Karaarslan E, Kural AR, Oto A, Purysko AS, Antic T, Magi-Galluzzi C, Saglican Y, Sioletic S, Warren AY, Bittencourt L, Fütterer JJ, Gupta RT, Kabakus I, Law YM, Margolis DJ, Shebel H, Westphalen AC, Wood BJ, Pinto PA, Shih JH, Choyke PL, Summers RM, Turkbey B. Can computer-aided diagnosis assist in the identification of prostate cancer on prostate MRI? a multi-center, multi-reader investigation. Oncotarget 2018; 9:33804-33817. [PMID: 30333911 PMCID: PMC6173466 DOI: 10.18632/oncotarget.26100] [Citation(s) in RCA: 59] [Impact Index Per Article: 9.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2018] [Accepted: 08/23/2018] [Indexed: 12/31/2022] Open
Abstract
For prostate cancer detection on prostate multiparametric MRI (mpMRI), the Prostate Imaging-Reporting and Data System version 2 (PI-RADSv2) and computer-aided diagnosis (CAD) systems aim to widely improve standardization across radiologists and centers. Our goal was to evaluate CAD assistance in prostate cancer detection compared with conventional mpMRI interpretation in a diverse dataset acquired from five institutions tested by nine readers of varying experience levels, in total representing 14 globally spread institutions. Index lesion sensitivities of mpMRI-alone were 79% (whole prostate (WP)), 84% (peripheral zone (PZ)), 71% (transition zone (TZ)), similar to CAD at 76% (WP, p=0.39), 77% (PZ, p=0.07), 79% (TZ, p=0.15). Greatest CAD benefit was in TZ for moderately-experienced readers at PI-RADSv2 <3 (84% vs mpMRI-alone 67%, p=0.055). Detection agreement was unchanged but CAD-assisted read times improved (4.6 vs 3.4 minutes, p<0.001). At PI-RADSv2 ≥ 3, CAD improved patient-level specificity (72%) compared to mpMRI-alone (45%, p<0.001). PI-RADSv2 and CAD-assisted mpMRI interpretations have similar sensitivities across multiple sites and readers while CAD has potential to improve specificity and moderately-experienced radiologists' detection of more difficult tumors in the center of the gland. The multi-institutional evidence provided is essential to future prostate MRI and CAD development.
Collapse
Affiliation(s)
- Sonia Gaur
- Molecular Imaging Program, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA
| | - Nathan Lay
- Imaging Biomarkers and Computer-aided Diagnosis Lab, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, MD, USA
| | - Stephanie A. Harmon
- Molecular Imaging Program, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA
- Clinical Research Directorate/ Clinical Monitoring Research Program, Leidos Biomedical Research, Inc., Frederick National Laboratory for Cancer Research, Frederick, MD, USA
| | - Sreya Doddakashi
- Molecular Imaging Program, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA
| | - Sherif Mehralivand
- Molecular Imaging Program, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA
- Urologic Oncology Branch, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA
- Department of Urology and Pediatric Urology, University Medical Center Mainz, Mainz, Germany
| | - Burak Argun
- Department of Urology, Acibadem University, Istanbul, Turkey
| | - Tristan Barrett
- Department of Radiology, University of Cambridge, Cambridge, UK
| | | | | | | | - Ali Riza Kural
- Department of Urology, Acibadem University, Istanbul, Turkey
| | - Aytekin Oto
- Department of Radiology, University of Chicago, Chicago, IL, USA
| | | | - Tatjana Antic
- Department of Pathology, University of Chicago, Chicago, IL, USA
| | | | - Yesim Saglican
- Department of Pathology, Acibadem University, Istanbul, Turkey
| | | | - Anne Y. Warren
- Department of Pathology, University of Cambridge, Cambridge, UK
| | | | | | - Rajan T. Gupta
- Department of Radiology, Duke University, Durham, NC, USA
| | - Ismail Kabakus
- Department of Radiology, Hacettepe University, Ankara, Turkey
| | - Yan Mee Law
- Department of Radiology, Singapore General Hospital, Singapore
| | | | - Haytham Shebel
- Department of Radiology, Mansoura University, Mansoura, Egypt
| | - Antonio C. Westphalen
- UCSF Department of Radiology, University of California-San Francisco, San Francisco, CA, USA
| | - Bradford J. Wood
- Center for Interventional Oncology, Clinical Center, National Institutes of Health, Bethesda, MD, USA
| | - Peter A. Pinto
- Urologic Oncology Branch, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA
| | - Joanna H. Shih
- Biometric Research Branch, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA
| | - Peter L. Choyke
- Molecular Imaging Program, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA
| | - Ronald M. Summers
- Imaging Biomarkers and Computer-aided Diagnosis Lab, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, MD, USA
| | - Baris Turkbey
- Molecular Imaging Program, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA
| |
Collapse
|
24
|
Takahashi R, Kajikawa Y. Computer-aided diagnosis: A survey with bibliometric analysis. Int J Med Inform 2017; 101:58-67. [DOI: 10.1016/j.ijmedinf.2017.02.004] [Citation(s) in RCA: 33] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2016] [Revised: 01/28/2017] [Accepted: 02/04/2017] [Indexed: 12/18/2022]
|