1
|
Abbasi N, Lacson R, Kapoor N, Licaros A, Guenette JP, Burk KS, Hammer M, Desai S, Eappen S, Saini S, Khorasani R. Development and External Validation of an Artificial Intelligence Model for Identifying Radiology Reports Containing Recommendations for Additional Imaging. AJR Am J Roentgenol 2023; 221:377-385. [PMID: 37073901 DOI: 10.2214/ajr.23.29120] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/20/2023]
Abstract
BACKGROUND. Reported rates of recommendations for additional imaging (RAIs) in radiology reports are low. Bidirectional encoder representations from transformers (BERT), a deep learning model pretrained to understand language context and ambiguity, has potential for identifying RAIs and thereby assisting large-scale quality improvement efforts. OBJECTIVE. The purpose of this study was to develop and externally validate an artificial intelligence (AI)-based model for identifying radiology reports containing RAIs. METHODS. This retrospective study was performed at a multisite health center. A total of 6300 radiology reports generated at one site from January 1, 2015, to June 30, 2021, were randomly selected and split by 4:1 ratio to create training (n = 5040) and test (n = 1260) sets. A total of 1260 reports generated at the center's other sites (including academic and community hospitals) from April 1 to April 30, 2022, were randomly selected as an external validation group. Referring practitioners and radiologists of varying sub-specialties manually reviewed report impressions for presence of RAIs. A BERT-based technique for identifying RAIs was developed by use of the training set. Performance of the BERT-based model and a previously developed traditional machine learning (TML) model was assessed in the test set. Finally, performance was assessed in the external validation set. The code for the BERT-based RAI model is publicly available. RESULTS. Among a total of 7419 unique patients (4133 women, 3286 men; mean age, 58.8 years), 10.0% of 7560 reports contained RAI. In the test set, the BERT-based model had 94.4% precision, 98.5% recall, and an F1 score of 96.4%. In the test set, the TML model had 69.0% precision, 65.4% recall, and an F1 score of 67.2%. In the test set, accuracy was greater for the BERT-based than for the TML model (99.2% vs 93.1%, p < .001). In the external validation set, the BERT-based model had 99.2% precision, 91.6% recall, an F1 score of 95.2%, and 99.0% accuracy. CONCLUSION. The BERT-based AI model accurately identified reports with RAIs, outperforming the TML model. High performance in the external validation set suggests the potential for other health systems to adapt the model without requiring institution-specific training. CLINICAL IMPACT. The model could potentially be used for real-time EHR monitoring for RAIs and other improvement initiatives to help ensure timely performance of clinically necessary recommended follow-up.
Collapse
Affiliation(s)
- Nooshin Abbasi
- Department of Radiology, Center for Evidence-Based Imaging, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
| | - Ronilda Lacson
- Department of Radiology, Brigham and Women's Hospital, Harvard Medical School, 75 Francis St, Boston, MA 02115
| | - Neena Kapoor
- Department of Radiology, Center for Evidence-Based Imaging, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
- Department of Radiology, Brigham and Women's Hospital, Harvard Medical School, 75 Francis St, Boston, MA 02115
| | - Andro Licaros
- Department of Radiology, Center for Evidence-Based Imaging, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
| | - Jeffrey P Guenette
- Department of Radiology, Center for Evidence-Based Imaging, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
- Department of Radiology, Brigham and Women's Hospital, Harvard Medical School, 75 Francis St, Boston, MA 02115
| | - Kristine Specht Burk
- Department of Radiology, Center for Evidence-Based Imaging, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
- Department of Radiology, Brigham and Women's Hospital, Harvard Medical School, 75 Francis St, Boston, MA 02115
| | - Mark Hammer
- Department of Radiology, Center for Evidence-Based Imaging, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
- Department of Radiology, Brigham and Women's Hospital, Harvard Medical School, 75 Francis St, Boston, MA 02115
| | - Sonali Desai
- Department of Medicine, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
| | - Sunil Eappen
- Department of Anesthesiology, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
| | - Sanjay Saini
- Department of Radiology, Center for Evidence-Based Imaging, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
- Department of Radiology, Massachusetts General Hospital, Harvard Medical School, Boston, MA
| | - Ramin Khorasani
- Department of Radiology, Center for Evidence-Based Imaging, Brigham and Women's Hospital, Harvard Medical School, Boston, MA
- Department of Radiology, Brigham and Women's Hospital, Harvard Medical School, 75 Francis St, Boston, MA 02115
| |
Collapse
|
2
|
Artificial Intelligence (AI) for Lung Nodules, From the AJR Special Series on AI Applications. AJR Am J Roentgenol 2022; 219:703-712. [PMID: 35544377 DOI: 10.2214/ajr.22.27487] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
Abstract
Interest in artificial intelligence (AI) applications for lung nodules continues to grow among radiologists, particularly with the expanding eligibility criteria and clinical utilization of lung cancer screening CT. AI has been heavily investigated for detecting and characterizing lung nodules and for guiding prognostic assessment. AI tools have also been used for image postprocessing (e.g., rib suppression on radiography or vessel suppression on CT) and for noninterpretive aspects of reporting and workflow, including management of nodule follow-up. Despite growing interest in and rapid development of AI tools and FDA approval of AI tools for pulmonary nodule evaluation, integration into clinical practice has been limited. Challenges to clinical adoption have included concerns about generalizability, regulatory issues, technical hurdles in implementation, and human skepticism. Further validation of AI tools for clinical use and demonstration of benefit in terms of patient-oriented outcomes also are needed. This article provides an overview of potential applications of AI tools in the imaging evaluation of lung nodules and discusses the challenges faced by practices interested in clinical implementation of such tools.
Collapse
|