1
|
Postiglione TJ, Guillo E, Heraud A, Rossillon A, Bartoli M, Herpe G, Adam C, Fabre D, Ardon R, Azarine A, Haulon S. Multicentric clinical evaluation of a computed tomography-based fully automated deep neural network for aortic maximum diameter and volumetric measurements. J Vasc Surg 2024; 79:1390-1400.e8. [PMID: 38325564 DOI: 10.1016/j.jvs.2024.01.214] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2023] [Revised: 01/04/2024] [Accepted: 01/04/2024] [Indexed: 02/09/2024]
Abstract
OBJECTIVE This study aims to evaluate a fully automatic deep learning-based method (augmented radiology for vascular aneurysm [ARVA]) for aortic segmentation and simultaneous diameter and volume measurements. METHODS A clinical validation dataset was constructed from preoperative and postoperative aortic computed tomography angiography (CTA) scans for assessing these functions. The dataset totaled 350 computed tomography angiography scans from 216 patients treated at two different hospitals. ARVA's ability to segment the aorta into seven morphologically based aortic segments and measure maximum outer-to-outer wall transverse diameters and compute volumes for each was compared with the measurements of six experts (ground truth) and thirteen clinicians. RESULTS Ground truth (experts') measurements of diameters and volumes were manually performed for all aortic segments. The median absolute diameter difference between ground truth and ARVA was 1.6 mm (95% confidence interval [CI], 1.5-1.7; and 1.6 mm [95% CI, 1.6-1.7]) between ground truth and clinicians. ARVA produced measurements within the clinical acceptable range with a proportion of 85.5% (95% CI, 83.5-86.3) compared with the clinicians' 86.0% (95% CI, 83.9-86.0). The median volume similarity error ranged from 0.93 to 0.95 in the main trunk and achieved 0.88 in the iliac arteries. CONCLUSIONS This study demonstrates the reliability of a fully automated artificial intelligence-driven solution capable of quick aortic segmentation and analysis of both diameter and volume for each segment.
Collapse
Affiliation(s)
- Thomas J Postiglione
- Aortic Centre, Hôpital Marie Lannelongue, Groupe Hospitalier Paris Saint Joseph, Université Paris Saclay, Paris, France
| | - Enora Guillo
- Radiology Department, Groupe Hospitalier Paris Saint Joseph, Paris, France
| | - Alexandre Heraud
- Radiology Department, Groupe Hospitalier Paris Saint Joseph, Paris, France
| | | | | | - Guillaume Herpe
- DACTIM MIS Lab, I3M, CNRS UMR, Poitiers, France; Incepto Medical, Paris, France
| | | | - Dominique Fabre
- Aortic Centre, Hôpital Marie Lannelongue, Groupe Hospitalier Paris Saint Joseph, Université Paris Saclay, Paris, France
| | | | - Arshid Azarine
- Radiology Department, Groupe Hospitalier Paris Saint Joseph, Paris, France
| | - Stéphan Haulon
- Aortic Centre, Hôpital Marie Lannelongue, Groupe Hospitalier Paris Saint Joseph, Université Paris Saclay, Paris, France.
| |
Collapse
|
2
|
Zhang Y, Joshi J, Hadi M. AI in Acute Cerebrovascular Disorders: What can the Radiologist Contribute? Semin Roentgenol 2024; 59:137-147. [PMID: 38880512 DOI: 10.1053/j.ro.2024.01.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2023] [Revised: 01/13/2024] [Accepted: 01/27/2024] [Indexed: 06/18/2024]
Affiliation(s)
- Yi Zhang
- Department of Radiology, University of Louisville, 530 South Jackson Street, CCB-C07, Louisville, KY
| | - Jonathan Joshi
- Department of Radiology, University of Louisville, 530 South Jackson Street, CCB-C07, Louisville, KY
| | - Mohiuddin Hadi
- Department of Radiology, University of Louisville, 530 South Jackson Street, CCB-C07, Louisville, KY.
| |
Collapse
|
3
|
Zhong J, Xing Y, Lu J, Zhang G, Mao S, Chen H, Yin Q, Cen Q, Jiang R, Hu Y, Ding D, Ge X, Zhang H, Yao W. The endorsement of general and artificial intelligence reporting guidelines in radiological journals: a meta-research study. BMC Med Res Methodol 2023; 23:292. [PMID: 38093215 PMCID: PMC10717715 DOI: 10.1186/s12874-023-02117-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Accepted: 12/01/2023] [Indexed: 12/17/2023] Open
Abstract
BACKGROUND Complete reporting is essential for clinical research. However, the endorsement of reporting guidelines in radiological journals is still unclear. Further, as a field extensively utilizing artificial intelligence (AI), the adoption of both general and AI reporting guidelines would be necessary for enhancing quality and transparency of radiological research. This study aims to investigate the endorsement of general reporting guidelines and those for AI applications in medical imaging in radiological journals, and explore associated journal characteristic variables. METHODS This meta-research study screened journals from the Radiology, Nuclear Medicine & Medical Imaging category, Science Citation Index Expanded of the 2022 Journal Citation Reports, and excluded journals not publishing original research, in non-English languages, and instructions for authors unavailable. The endorsement of fifteen general reporting guidelines and ten AI reporting guidelines was rated using a five-level tool: "active strong", "active weak", "passive moderate", "passive weak", and "none". The association between endorsement and journal characteristic variables was evaluated by logistic regression analysis. RESULTS We included 117 journals. The top-five endorsed reporting guidelines were CONSORT (Consolidated Standards of Reporting Trials, 58.1%, 68/117), PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses, 54.7%, 64/117), STROBE (STrengthening the Reporting of Observational Studies in Epidemiology, 51.3%, 60/117), STARD (Standards for Reporting of Diagnostic Accuracy, 50.4%, 59/117), and ARRIVE (Animal Research Reporting of In Vivo Experiments, 35.9%, 42/117). The most implemented AI reporting guideline was CLAIM (Checklist for Artificial Intelligence in Medical Imaging, 1.7%, 2/117), while other nine AI reporting guidelines were not mentioned. The Journal Impact Factor quartile and publisher were associated with endorsement of reporting guidelines in radiological journals. CONCLUSIONS The general reporting guideline endorsement was suboptimal in radiological journals. The implementation of reporting guidelines for AI applications in medical imaging was extremely low. Their adoption should be strengthened to facilitate quality and transparency of radiological study reporting.
Collapse
Affiliation(s)
- Jingyu Zhong
- Department of Imaging, Tongren Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200336, China
| | - Yue Xing
- Department of Imaging, Tongren Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200336, China
| | - Junjie Lu
- Department of Epidemiology and Population Health, Stanford University School of Medicine, Stanford, CA, 94305, USA
| | - Guangcheng Zhang
- Department of Orthopedics, Shanghai Sixth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200233, China
| | - Shiqi Mao
- Department of Medical Oncology, Shanghai Pulmonary Hospital, Tongji University School of Medicine, Shanghai, 200433, China
| | - Haoda Chen
- Department of General Surgery, Pancreatic Disease Center, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200025, China
| | - Qian Yin
- Department of Pathology, Shanghai Sixth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200233, China
| | - Qingqing Cen
- Department of Dermatology, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200011, China
| | - Run Jiang
- Department of Pharmacovigilance, Shanghai Hansoh BioMedical Co., Ltd., Shanghai, 201203, China
| | - Yangfan Hu
- Department of Imaging, Tongren Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200336, China
| | - Defang Ding
- Department of Imaging, Tongren Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200336, China
| | - Xiang Ge
- Department of Imaging, Tongren Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200336, China
| | - Huan Zhang
- Department of Radiology, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200025, China.
| | - Weiwu Yao
- Department of Imaging, Tongren Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, 200336, China.
| |
Collapse
|
4
|
Lee KH, Lee RW, Lee KH, Park W, Kwon SR, Lim MJ. The Development and Validation of an AI Diagnostic Model for Sacroiliitis: A Deep-Learning Approach. Diagnostics (Basel) 2023; 13:3643. [PMID: 38132228 PMCID: PMC10743277 DOI: 10.3390/diagnostics13243643] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2023] [Revised: 12/07/2023] [Accepted: 12/11/2023] [Indexed: 12/23/2023] Open
Abstract
PURPOSE Sacroiliitis refers to the inflammatory condition of the sacroiliac joints, frequently causing lower back pain. It is often associated with systemic conditions. However, its signs on radiographic images can be subtle, which may result in it being overlooked or underdiagnosed. This study aims to utilize artificial intelligence (AI) to create a diagnostic tool for more accurate sacroiliitis detection in radiological images, with the goal of optimizing treatment plans and improving patient outcomes. MATERIALS AND METHOD The study included 492 patients who visited our hospital. Right sacroiliac joint films were independently evaluated by two musculoskeletal radiologists using the Modified New York criteria (Normal, Grades 1-4). A consensus reading resolved disagreements. The images were preprocessed with Z-score standardization and histogram equalization. The DenseNet121 algorithm, a convolutional neural network with 201 layers, was used for learning and classification. All steps were performed on the DEEP:PHI platform. RESULT The AI model exhibited high accuracy across different grades: 94.53% (Grade 1), 95.83% (Grade 2), 98.44% (Grade 3), 96.88% (Grade 4), and 96.09% (Normal cases). Sensitivity peaked at Grade 3 and Normal cases (100%), while Grade 4 achieved perfect specificity (100%). PPVs ranged from 82.61% (Grade 1) to 100% (Grade 4), and NPVs peaked at 100% for Grade 3 and Normal cases. The F1 scores ranged from 64.41% (Grade 1) to 95.38% (Grade 3). CONCLUSIONS The AI diagnostic model showcased a robust performance in detecting and grading sacroiliitis, reflecting its potential to enhance diagnostic accuracy in clinical settings. By facilitating earlier and more accurate diagnoses, this model could substantially impact treatment strategies and patient outcomes.
Collapse
Affiliation(s)
- Kyu-Hong Lee
- Department of Radiology, College of Medicine, Inha University, Incheon 22212, Republic of Korea; (K.-H.L.)
| | - Ro-Woon Lee
- Department of Radiology, College of Medicine, Inha University, Incheon 22212, Republic of Korea; (K.-H.L.)
| | - Kyung-Hee Lee
- Department of Radiology, College of Medicine, Inha University, Incheon 22212, Republic of Korea; (K.-H.L.)
| | - Won Park
- Division of Rheumatology, Department of Internal Medicine, College of Medicine, Inha University, Incheon 22212, Republic of Korea; (W.P.)
| | - Seong-Ryul Kwon
- Division of Rheumatology, Department of Internal Medicine, College of Medicine, Inha University, Incheon 22212, Republic of Korea; (W.P.)
| | - Mie-Jin Lim
- Division of Rheumatology, Department of Internal Medicine, College of Medicine, Inha University, Incheon 22212, Republic of Korea; (W.P.)
| |
Collapse
|
5
|
Winter PD, Carusi A. (De)troubling transparency: artificial intelligence (AI) for clinical applications. MEDICAL HUMANITIES 2023; 49:17-26. [PMID: 35545432 PMCID: PMC9985768 DOI: 10.1136/medhum-2021-012318] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 04/05/2022] [Indexed: 06/15/2023]
Abstract
Artificial intelligence (AI) and machine learning (ML) techniques occupy a prominent role in medical research in terms of the innovation and development of new technologies. However, while many perceive AI as a technology of promise and hope-one that is allowing for more early and accurate diagnosis-the acceptance of AI and ML technologies in hospitals remains low. A major reason for this is the lack of transparency associated with these technologies, in particular epistemic transparency, which results in AI disturbing or troubling established knowledge practices in clinical contexts. In this article, we describe the development process of one AI application for a clinical setting. We show how epistemic transparency is negotiated and co-produced in close collaboration between AI developers and clinicians and biomedical scientists, forming the context in which AI is accepted as an epistemic operator. Drawing on qualitative research with collaborative researchers developing an AI technology for the early diagnosis of a rare respiratory disease (pulmonary hypertension/PH), this paper examines how including clinicians and clinical scientists in the collaborative practices of AI developers de-troubles transparency. Our research shows how de-troubling transparency occurs in three dimensions of AI development relating to PH: querying of data sets, building software and training the model The close collaboration results in an AI application that is at once social and technological: it integrates and inscribes into the technology the knowledge processes of the different participants in its development. We suggest that it is a misnomer to call these applications 'artificial' intelligence, and that they would be better developed and implemented if they were reframed as forms of sociotechnical intelligence.
Collapse
Affiliation(s)
- Peter David Winter
- School of Sociology, Politics and International Studies, University of Bristol, Bristol, UK
| | - Annamaria Carusi
- Interchange Research, London, UK
- Department of Science and Technology Studies, University College London, London, London, UK
| |
Collapse
|
6
|
Carusi A, Winter PD, Armstrong I, Ciravegna F, Kiely DG, Lawrie A, Lu H, Sabroe I, Swift A. Medical artificial intelligence is as much social as it is technological. NAT MACH INTELL 2023. [DOI: 10.1038/s42256-022-00603-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
|
7
|
Bangert P, Moon H, Woo JO, Didari S, Hao H. Active Learning Performance in Labeling Radiology Images Is 90% Effective. FRONTIERS IN RADIOLOGY 2021; 1:748968. [PMID: 37492167 PMCID: PMC10365082 DOI: 10.3389/fradi.2021.748968] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/28/2021] [Accepted: 10/28/2021] [Indexed: 07/27/2023]
Abstract
To train artificial intelligence (AI) systems on radiology images, an image labeling step is necessary. Labeling for radiology images usually involves a human radiologist manually drawing a (polygonal) shape onto the image and attaching a word to it. As datasets are typically large, this task is repetitive, time-consuming, error-prone, and expensive. The AI methodology of active learning (AL) can assist human labelers by continuously sorting the unlabeled images in order of information gain and thus getting the labeler always to label the most informative image next. We find that after about 10%, depending on the dataset, of the images in a realistic dataset are labeled, virtually all the information content has been learnt and the remaining images can be automatically labeled. These images can then be checked by the radiologist, which is far easier and faster to do. In this way, the entire dataset is labeled with much less human effort. We introduce AL in detail and expose the effectiveness using three real-life datasets. We contribute five distinct elements to the standard AL workflow creating an advanced methodology.
Collapse
|