1
|
Jiménez-Gaona Y, Álvarez MJR, Castillo-Malla D, García-Jaen S, Carrión-Figueroa D, Corral-Domínguez P, Lakshminarayanan V. BraNet: a mobil application for breast image classification based on deep learning algorithms. Med Biol Eng Comput 2024; 62:2737-2756. [PMID: 38693328 PMCID: PMC11330402 DOI: 10.1007/s11517-024-03084-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2023] [Accepted: 03/26/2024] [Indexed: 05/03/2024]
Abstract
Mobile health apps are widely used for breast cancer detection using artificial intelligence algorithms, providing radiologists with second opinions and reducing false diagnoses. This study aims to develop an open-source mobile app named "BraNet" for 2D breast imaging segmentation and classification using deep learning algorithms. During the phase off-line, an SNGAN model was previously trained for synthetic image generation, and subsequently, these images were used to pre-trained SAM and ResNet18 segmentation and classification models. During phase online, the BraNet app was developed using the react native framework, offering a modular deep-learning pipeline for mammography (DM) and ultrasound (US) breast imaging classification. This application operates on a client-server architecture and was implemented in Python for iOS and Android devices. Then, two diagnostic radiologists were given a reading test of 290 total original RoI images to assign the perceived breast tissue type. The reader's agreement was assessed using the kappa coefficient. The BraNet App Mobil exhibited the highest accuracy in benign and malignant US images (94.7%/93.6%) classification compared to DM during training I (80.9%/76.9%) and training II (73.7/72.3%). The information contrasts with radiological experts' accuracy, with DM classification being 29%, concerning US 70% for both readers, because they achieved a higher accuracy in US ROI classification than DM images. The kappa value indicates a fair agreement (0.3) for DM images and moderate agreement (0.4) for US images in both readers. It means that not only the amount of data is essential in training deep learning algorithms. Also, it is vital to consider the variety of abnormalities, especially in the mammography data, where several BI-RADS categories are present (microcalcifications, nodules, mass, asymmetry, and dense breasts) and can affect the API accuracy model.
Collapse
Affiliation(s)
- Yuliana Jiménez-Gaona
- Departamento de Química y Ciencias Exactas, Universidad Técnica Particular de Loja, San Cayetano Alto s/n CP1101608, Loja, Ecuador.
- Instituto de Instrumentación para la Imagen Molecular I3M, Universitat Politécnica de Valencia, 46022, Valencia, Spain.
- Theoretical and Experimental Epistemology Lab, School of Opto ΩN2L3G1, Waterloo, Canada.
| | - María José Rodríguez Álvarez
- Instituto de Instrumentación para la Imagen Molecular I3M, Universitat Politécnica de Valencia, 46022, Valencia, Spain
| | - Darwin Castillo-Malla
- Departamento de Química y Ciencias Exactas, Universidad Técnica Particular de Loja, San Cayetano Alto s/n CP1101608, Loja, Ecuador
- Instituto de Instrumentación para la Imagen Molecular I3M, Universitat Politécnica de Valencia, 46022, Valencia, Spain
- Theoretical and Experimental Epistemology Lab, School of Opto ΩN2L3G1, Waterloo, Canada
| | - Santiago García-Jaen
- Departamento de Química y Ciencias Exactas, Universidad Técnica Particular de Loja, San Cayetano Alto s/n CP1101608, Loja, Ecuador
| | | | - Patricio Corral-Domínguez
- Corporación Médica Monte Sinaí-CIPAM (Centro Integral de Patología Mamaria) Cuenca-Ecuador, Facultad de Ciencias Médicas, Universidad de Cuenca, Cuenca, 010203, Ecuador
| | - Vasudevan Lakshminarayanan
- Department of Systems Design Engineering, Physics, and Electrical and Computer Engineering, University of Waterloo, Waterloo, ON, N2L3G1, Canada
| |
Collapse
|
2
|
Sastry RA, Setty A, Liu DD, Zheng B, Ali R, Weil RJ, Roye GD, Doberstein CE, Oyelese AA, Niu T, Gokaslan ZL, Telfeian AE. Natural language processing augments comorbidity documentation in neurosurgical inpatient admissions. PLoS One 2024; 19:e0303519. [PMID: 38723044 PMCID: PMC11081267 DOI: 10.1371/journal.pone.0303519] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2022] [Accepted: 04/04/2024] [Indexed: 05/13/2024] Open
Abstract
OBJECTIVE To establish whether or not a natural language processing technique could identify two common inpatient neurosurgical comorbidities using only text reports of inpatient head imaging. MATERIALS AND METHODS A training and testing dataset of reports of 979 CT or MRI scans of the brain for patients admitted to the neurosurgery service of a single hospital in June 2021 or to the Emergency Department between July 1-8, 2021, was identified. A variety of machine learning and deep learning algorithms utilizing natural language processing were trained on the training set (84% of the total cohort) and tested on the remaining images. A subset comparison cohort (n = 76) was then assessed to compare output of the best algorithm against real-life inpatient documentation. RESULTS For "brain compression", a random forest classifier outperformed other candidate algorithms with an accuracy of 0.81 and area under the curve of 0.90 in the testing dataset. For "brain edema", a random forest classifier again outperformed other candidate algorithms with an accuracy of 0.92 and AUC of 0.94 in the testing dataset. In the provider comparison dataset, for "brain compression," the random forest algorithm demonstrated better accuracy (0.76 vs 0.70) and sensitivity (0.73 vs 0.43) than provider documentation. For "brain edema," the algorithm again demonstrated better accuracy (0.92 vs 0.84) and AUC (0.45 vs 0.09) than provider documentation. DISCUSSION A natural language processing-based machine learning algorithm can reliably and reproducibly identify selected common neurosurgical comorbidities from radiology reports. CONCLUSION This result may justify the use of machine learning-based decision support to augment provider documentation.
Collapse
Affiliation(s)
- Rahul A. Sastry
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| | - Aayush Setty
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
- Department of Computer Science, Brown University, Providence, RI, United States of America
| | - David D. Liu
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| | - Bryan Zheng
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| | - Rohaid Ali
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| | - Robert J. Weil
- Department of Neurosurgery, Brain & Spine, Southcoast Health, Dartmouth, MA, United States of America
| | - G. Dean Roye
- Department of Surgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| | - Curtis E. Doberstein
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| | - Adetokunbo A. Oyelese
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| | - Tianyi Niu
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| | - Ziya L. Gokaslan
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| | - Albert E. Telfeian
- Department of Neurosurgery, Warren Alpert Medical School, Rhode Island Hospital, Brown University, Providence, RI, United States of America
| |
Collapse
|
3
|
Khoruzhaya A, Kozlov D, Arzamasov K, Kremneva E. Comparison of an Ensemble of Machine Learning Models and the BERT Language Model for Analysis of Text Descriptions of Brain CT Reports to Determine the Presence of Intracranial Hemorrhage. Sovrem Tekhnologii Med 2024; 16:27-34. [PMID: 39421632 PMCID: PMC11482096 DOI: 10.17691/stm2024.16.1.03] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2023] [Indexed: 10/19/2024] Open
Abstract
The aim of this study is to train and test an ensemble of machine learning models, as well as to compare its performance with the BERT language model pre-trained on medical data to perform simple binary classification, i.e., determine the presence/absence of the signs of intracranial hemorrhage (ICH) in brain CT reports. Materials and Methods Seven machine learning algorithms and three text vectorization techniques were selected as models to solve the binary classification problem. These models were trained on textual data represented by 3980 brain CT reports from 56 inpatient medical facilities in Moscow. The study utilized three text vectorization techniques: bag of words, TF-IDF, and word2vec. The resulting data were then processed by the following machine learning algorithms: decision tree, random forest, logistic regression, nearest neighbors, support vector machines, Catboost, and XGboost. Data analysis and pre-processing were performed using NLTK (Natural Language Toolkit, version 3.6.5), libraries for character-based and statistical processing of natural language, and Scikit-learn (version 0.24.2), a library for machine learning containing tools to tackle classification challenges. MedRuBertTiny2 was taken as a BERT transformer model pre-trained on medical data. Results Based on the training and testing outcomes from seven machine learning algorithms, the authors selected three algorithms that yielded the highest metrics (i.e. sensitivity and specificity): CatBoost, logistic regression, and nearest neighbors. The highest metrics were achieved by the bag of words technique. These algorithms were assembled into an ensemble using the stacking technique. The sensitivity and specificity for the validation dataset separated from the original sample were 0.93 and 0.90, respectively. Next, the ensemble and the BERT model were trained on an independent dataset containing 9393 textual radiology reports also divided into training and test sets. Once the ensemble was tested on this dataset, the resulting sensitivity and specificity were 0.92 and 0.90, respectively. The BERT model tested on these data demonstrated a sensitivity of 0.97 and a specificity of 0.90. Conclusion When analyzing textual reports of brain CT scans with signs of intracranial hemorrhage, the trained ensemble demonstrated high accuracy metrics. Still, manual quality control of the results is required during its application. The pre-trained BERT transformer model, additionally trained on diagnostic textual reports, demonstrated higher accuracy metrics (p<0.05). The results show promise in terms of finding specific values for both binary classification task and in-depth analysis of unstructured medical information.
Collapse
Affiliation(s)
- A.N. Khoruzhaya
- Junior Researcher, Department of Innovative Technologies; Scientific and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Department of Health, Bldg 1, 24 Petrovka St., Moscow, 127051, Russia
| | - D.V. Kozlov
- Junior Researcher, Department of Medical Informatics, Radiomics and Radiogenomics; Scientific and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Department of Health, Bldg 1, 24 Petrovka St., Moscow, 127051, Russia
| | - K.M. Arzamasov
- Head of the Department of Medical Informatics, Radiomics and Radiogenomics; Scientific and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Department of Health, Bldg 1, 24 Petrovka St., Moscow, 127051, Russia
| | - E.I. Kremneva
- Leading Researcher, Department of Innovative Technologies; Scientific and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Department of Health, Bldg 1, 24 Petrovka St., Moscow, 127051, Russia; Senior Researcher; Research Center for Neurology, 80 Volokolamskoye Shosse, Moscow, 125367, Russia
| |
Collapse
|
4
|
Basilio R, Carvalho AR, Rodrigues R, Conrado M, Accorsi S, Forghani R, Machuca T, Zanon M, Altmayer S, Hochhegger B. Natural Language Processing for the Identification of Incidental Lung Nodules in Computed Tomography Reports: A Quality Control Tool. JCO Glob Oncol 2023; 9:e2300191. [PMID: 37769221 PMCID: PMC10581645 DOI: 10.1200/go.23.00191] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Revised: 07/09/2023] [Accepted: 08/22/2023] [Indexed: 09/30/2023] Open
Abstract
PURPOSE To evaluate the diagnostic performance of a natural language processing (NLP) model in detecting incidental lung nodules (ILNs) in unstructured chest computed tomography (CT) reports. METHODS All unstructured consecutive reports of chest CT scans performed at a tertiary hospital between 2020 and 2021 were retrospectively reviewed (n = 21,542) to train the NLP tool. Internal validation was performed using reference readings by two radiologists of both CT scans and reports, using a different external cohort of 300 chest CT scans. Second, external validation was performed in a cohort of all random unstructured chest CT reports from 57 different hospitals conducted in May 2022. A review by the same thoracic radiologists was used as the gold standard. The sensitivity, specificity, and accuracy were calculated. RESULTS Of 21,542 CT reports, 484 mentioned at least one ILN (mean age, 71 ± 17.6 [standard deviation] years; women, 52%) and were included in the training set. In the internal validation (n = 300), the NLP tool detected ILN with a sensitivity of 100.0% (95% CI, 97.6 to 100.0), a specificity of 95.9% (95% CI, 91.3 to 98.5), and an accuracy of 98.0% (95% CI, 95.7 to 99.3). In the external validation (n = 977), the NLP tool yielded a sensitivity of 98.4% (95% CI, 94.5 to 99.8), a specificity of 98.6% (95% CI, 97.5 to 99.3), and an accuracy of 98.6% (95% CI, 97.6 to 99.2). Twelve months after the initial reports, 8 (8.60%) patients had a final diagnosis of lung cancer, among which 2 (2.15%) would have been lost to follow-up without the NLP tool. CONCLUSION NLP can be used to identify ILNs in unstructured reports with high accuracy, allowing a timely recall of patients and a potential diagnosis of early-stage lung cancer that might have been lost to follow-up.
Collapse
Affiliation(s)
- Rodrigo Basilio
- D'Or Institute for Research and Education (IDOR), Rio de Janeiro, Brazil
| | | | - Rosana Rodrigues
- D'Or Institute for Research and Education (IDOR), Rio de Janeiro, Brazil
| | - Marco Conrado
- D'Or Institute for Research and Education (IDOR), Rio de Janeiro, Brazil
| | - Sephania Accorsi
- D'Or Institute for Research and Education (IDOR), Rio de Janeiro, Brazil
| | - Reza Forghani
- Radiomics and Augmented Intelligence Laboratory (RAIL), University of Florida, Gainesville, FL
| | - Tiago Machuca
- D'Or Institute for Research and Education (IDOR), Rio de Janeiro, Brazil
| | - Matheus Zanon
- Federal University of Health Sciences of Porto Alegre, Porto Alegre, Brazil
| | - Stephan Altmayer
- Stanford Hospital, Stanford University Medical Center, Palo Alto, CA
| | - Bruno Hochhegger
- Radiomics and Augmented Intelligence Laboratory (RAIL), University of Florida, Gainesville, FL
- Federal University of Health Sciences of Porto Alegre, Porto Alegre, Brazil
| |
Collapse
|
5
|
Ståhlbrandt H, Björnfot I, Cederlund T, Almén A. CT and MRI imaging in Sweden: retrospective appropriateness analysis of large referral samples. Insights Imaging 2023; 14:134. [PMID: 37530862 PMCID: PMC10397157 DOI: 10.1186/s13244-023-01483-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2023] [Accepted: 07/04/2023] [Indexed: 08/03/2023] Open
Abstract
OBJECTIVES The numbers of computed tomography (CT) and magnetic resonance imaging (MRI) examinations per capita continue to increase in Sweden and in other parts of Europe. The appropriateness of CT and MRI examinations was audited using established European appropriateness criteria. Alternative modalities were also explored. The results were compared with those of a previous study performed in Sweden. METHODS A semi-automatic retrospective evaluation of referrals from examinations performed in four healthcare regions using the European appropriateness criteria in ESR iGuide was undertaken. The clinical indications from a total of 13,075 referrals were assessed against these criteria. The ESR iGuide was used to identify alternative modalities resulting in a higher degree of appropriateness. A qualitative comparison with re-evaluated results from the previous study was made. RESULTS The appropriateness was higher for MRI examinations than for CT examinations with procedures classed as usually appropriate for 76% and 63% of the examinations, respectively. The degree of appropriateness for CT was higher for referrals from hospitals compared to those from primary care centres. The opposite was found for MRI examinations. The alternative modalities that would result in higher appropriateness included all main imaging modalities. The result for CT did not show improvement compared with the former study. CONCLUSIONS A high proportion of both CT and MRI examinations were inappropriate. The study indicates that 37% of CT examinations and 24% of MRI examinations were inappropriate and that the appropriateness for CT has not improved in the last 15 years. CRITICAL RELEVANCE STATEMENT A high proportion of CT and MRI examinations in this retrospective study using evidence-based referral guidelines were inappropriate. KEY POINTS ∙ A high proportion of CT and MRI examinations were inappropriate. ∙ The CT referrals from general practitioners were less appropriate that those from hospital specialists. ∙ The MRI referrals from hospital specialists were less appropriate that those from general practitioners. ∙ Adherence to radiological appropriateness guidelines may improve the appropriateness of conducted examinations.
Collapse
Affiliation(s)
| | - Ida Björnfot
- Department of Radiology, Länssjukhuset Ryhov, 551 85, Jönköping, Sweden
| | - Torsten Cederlund
- Department for Authorisation of Radiation Applications, Swedish Radiation Safety Authority, 171 16, Stockholm, Sweden
| | - Anja Almén
- Department for Radiation Protection and Environmental Assessment, Swedish Radiation Safety Authority, 171 16, Stockholm, Sweden.
- Medical Radiation Physics, Department of Translational Medicine, Malmö, Lund University, 205 02, Malmö, Sweden.
| |
Collapse
|
6
|
Garcia EV. Integrating artificial intelligence and natural language processing for computer-assisted reporting and report understanding in nuclear cardiology. J Nucl Cardiol 2023; 30:1180-1190. [PMID: 35725887 DOI: 10.1007/s12350-022-02996-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2022] [Accepted: 03/29/2022] [Indexed: 10/18/2022]
Abstract
Natural language processing (NLP) offers many opportunities in Nuclear Cardiology. These opportunities include applications in converting nuclear cardiology imaging reports to digital searchable information that may be used as Big Data for machine learning and registries. Another major NLP application is, with the support of AI, in automatically translating MPI image features directly into nuclear cardiology reports. This review describes the symbiotic relationship between AI and NLP in that NLP is being used to facilitate AI applications and, AI techniques are being used to facilitate NLP. This article reviews the fundamentals of NLP and describes various conventional and AI techniques that have been applied in imaging. Key nuclear cardiology applications are reviewed such as conversion of MPI free-text reports to digital documents as well as direct conversion of MPI images into structured medical reports.
Collapse
Affiliation(s)
- Ernest V Garcia
- Department of Radiology and Imaging Sciences, Emory University School of Medicine, 101 Woodruff Circle, Room 1203, Atlanta, GA, 30322, USA.
| |
Collapse
|
7
|
Liu DS, Abu-Shaban K, Halabi SS, Cook TS. Changes in Radiology Due to Artificial Intelligence That Can Attract Medical Students to the Specialty. JMIR MEDICAL EDUCATION 2023; 9:e43415. [PMID: 36939823 PMCID: PMC10131993 DOI: 10.2196/43415] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/11/2022] [Revised: 02/19/2023] [Accepted: 02/25/2023] [Indexed: 06/18/2023]
Abstract
The role of artificial intelligence (AI) in radiology has grown exponentially in the recent years. One of the primary worries by medical students is that AI will cause the roles of a radiologist to become automated and thus obsolete. Therefore, there is a greater hesitancy by medical students to choose radiology as a specialty. However, it is in this time of change that the specialty needs new thinkers and leaders. In this succinct viewpoint, 2 medical students involved in AI and 2 radiologists specializing in AI or clinical informatics posit that not only are these fears false, but the field of radiology will be transformed in such a way due to AI that there will be novel reasons to choose radiology. These new factors include greater impact on patient care, new space for innovation, interdisciplinary collaboration, increased patient contact, becoming master diagnosticians, and greater opportunity for global health initiatives, among others. Finally, since medical students view mentorship as a critical resource when deciding their career path, medical educators must also be cognizant of these changes and not give much credence to the prevalent fearmongering. As the field and practice of radiology continue to undergo significant change due to AI, it is urgent and necessary for the conversation to expand from expert to expert to expert to student. Medical students should be encouraged to choose radiology specifically because of the changes brought on by AI rather than being deterred by it.
Collapse
Affiliation(s)
- David Shalom Liu
- University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
| | - Kamil Abu-Shaban
- University of Toledo College of Medicine and Life Sciences, Toledo, OH, United States
| | - Safwan S Halabi
- Department of Medical Imaging, Ann & Robert H Lurie Children's Hospital of Chicago, Chicago, IL, United States
| | - Tessa Sundaram Cook
- Department of Radiology, Hospital of the University of Pennsylvania, Pennsylvania, PA, United States
| |
Collapse
|
8
|
Natural Language Processing Applications for Computer-Aided Diagnosis in Oncology. Diagnostics (Basel) 2023; 13:diagnostics13020286. [PMID: 36673096 PMCID: PMC9857980 DOI: 10.3390/diagnostics13020286] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2022] [Revised: 12/24/2022] [Accepted: 01/05/2023] [Indexed: 01/15/2023] Open
Abstract
In the era of big data, text-based medical data, such as electronic health records (EHR) and electronic medical records (EMR), are growing rapidly. EHR and EMR are collected from patients to record their basic information, lab tests, vital signs, clinical notes, and reports. EHR and EMR contain the helpful information to assist oncologists in computer-aided diagnosis and decision making. However, it is time consuming for doctors to extract the valuable information they need and analyze the information from the EHR and EMR data. Recently, more and more research works have applied natural language processing (NLP) techniques, i.e., rule-based, machine learning-based, and deep learning-based techniques, on the EHR and EMR data for computer-aided diagnosis in oncology. The objective of this review is to narratively review the recent progress in the area of NLP applications for computer-aided diagnosis in oncology. Moreover, we intend to reduce the research gap between artificial intelligence (AI) experts and clinical specialists to design better NLP applications. We originally identified 295 articles from the three electronic databases: PubMed, Google Scholar, and ACL Anthology; then, we removed the duplicated papers and manually screened the irrelevant papers based on the content of the abstract; finally, we included a total of 23 articles after the screening process of the literature review. Furthermore, we provided an in-depth analysis and categorized these studies into seven cancer types: breast cancer, lung cancer, liver cancer, prostate cancer, pancreatic cancer, colorectal cancer, and brain tumors. Additionally, we identified the current limitations of NLP applications on supporting the clinical practices and we suggest some promising future research directions in this paper.
Collapse
|
9
|
Saha A, Burns L, Kulkarni AM. A scoping review of natural language processing of radiology reports in breast cancer. Front Oncol 2023; 13:1160167. [PMID: 37124523 PMCID: PMC10130381 DOI: 10.3389/fonc.2023.1160167] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2023] [Accepted: 03/28/2023] [Indexed: 05/02/2023] Open
Abstract
Various natural language processing (NLP) algorithms have been applied in the literature to analyze radiology reports pertaining to the diagnosis and subsequent care of cancer patients. Applications of this technology include cohort selection for clinical trials, population of large-scale data registries, and quality improvement in radiology workflows including mammography screening. This scoping review is the first to examine such applications in the specific context of breast cancer. Out of 210 identified articles initially, 44 met our inclusion criteria for this review. Extracted data elements included both clinical and technical details of studies that developed or evaluated NLP algorithms applied to free-text radiology reports of breast cancer. Our review illustrates an emphasis on applications in diagnostic and screening processes over treatment or therapeutic applications and describes growth in deep learning and transfer learning approaches in recent years, although rule-based approaches continue to be useful. Furthermore, we observe increased efforts in code and software sharing but not with data sharing.
Collapse
Affiliation(s)
- Ashirbani Saha
- Department of Oncology, McMaster University, Hamilton, ON, Canada
- Hamilton Health Sciences and McMaster University, Escarpment Cancer Research Institute, Hamilton, ON, Canada
- *Correspondence: Ashirbani Saha,
| | - Levi Burns
- Michael G. DeGroote School of Medicine, McMaster University, Hamilton, ON, Canada
| | | |
Collapse
|
10
|
Binsfeld Gonçalves L, Nesic I, Obradovic M, Stieltjes B, Weikert T, Bremerich J. Natural Language Processing and Graph Theory: Making Sense of Imaging Records in a Novel Representation Frame. JMIR Med Inform 2022; 10:e40534. [PMID: 36542426 PMCID: PMC9813822 DOI: 10.2196/40534] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2022] [Revised: 09/13/2022] [Accepted: 11/30/2022] [Indexed: 12/24/2022] Open
Abstract
BACKGROUND A concise visualization framework of related reports would increase readability and improve patient management. To this end, temporal referrals to prior comparative exams are an essential connection to previous exams in written reports. Due to unstructured narrative texts' variable structure and content, their extraction is hampered by poor computer readability. Natural language processing (NLP) permits the extraction of structured information from unstructured texts automatically and can serve as an essential input for such a novel visualization framework. OBJECTIVE This study proposes and evaluates an NLP-based algorithm capable of extracting the temporal referrals in written radiology reports, applies it to all the radiology reports generated for 10 years, introduces a graphical representation of imaging reports, and investigates its benefits for clinical and research purposes. METHODS In this single-center, university hospital, retrospective study, we developed a convolutional neural network capable of extracting the date of referrals from imaging reports. The model's performance was assessed by calculating precision, recall, and F1-score using an independent test set of 149 reports. Next, the algorithm was applied to our department's radiology reports generated from 2011 to 2021. Finally, the reports and their metadata were represented in a modulable graph. RESULTS For extracting the date of referrals, the named-entity recognition (NER) model had a high precision of 0.93, a recall of 0.95, and an F1-score of 0.94. A total of 1,684,635 reports were included in the analysis. Temporal reference was mentioned in 53.3% (656,852/1,684,635), explicitly stated as not available in 21.0% (258,386/1,684,635), and omitted in 25.7% (317,059/1,684,635) of the reports. Imaging records can be visualized in a directed and modulable graph, in which the referring links represent the connecting arrows. CONCLUSIONS Automatically extracting the date of referrals from unstructured radiology reports using deep learning NLP algorithms is feasible. Graphs refined the selection of distinct pathology pathways, facilitated the revelation of missing comparisons, and enabled the query of specific referring exam sequences. Further work is needed to evaluate its benefits in clinics, research, and resource planning.
Collapse
Affiliation(s)
- Laurent Binsfeld Gonçalves
- Clinic of Radiology & Nuclear Medicine, University Hospital Basel, University of Basel, Basel, Switzerland
| | - Ivan Nesic
- Clinic of Radiology & Nuclear Medicine, University Hospital Basel, University of Basel, Basel, Switzerland
| | - Marko Obradovic
- Clinic of Radiology & Nuclear Medicine, University Hospital Basel, University of Basel, Basel, Switzerland
| | - Bram Stieltjes
- Clinic of Radiology & Nuclear Medicine, University Hospital Basel, University of Basel, Basel, Switzerland
| | - Thomas Weikert
- Clinic of Radiology & Nuclear Medicine, University Hospital Basel, University of Basel, Basel, Switzerland
| | - Jens Bremerich
- Clinic of Radiology & Nuclear Medicine, University Hospital Basel, University of Basel, Basel, Switzerland
| |
Collapse
|
11
|
Horng H, Steinkamp J, Kahn CE, Cook TS. Ensemble Approaches to Recognize Protected Health Information in Radiology Reports. J Digit Imaging 2022; 35:1694-1698. [PMID: 35715655 PMCID: PMC9712864 DOI: 10.1007/s10278-022-00673-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2022] [Revised: 06/02/2022] [Accepted: 06/07/2022] [Indexed: 10/18/2022] Open
Abstract
Natural language processing (NLP) techniques for electronic health records have shown great potential to improve the quality of medical care. The text of radiology reports frequently constitutes a large fraction of EHR data, and can provide valuable information about patients' diagnoses, medical history, and imaging findings. The lack of a major public repository for radiological reports severely limits the development, testing, and application of new NLP tools. De-identification of protected health information (PHI) presents a major challenge to building such repositories, as many automated tools for de-identification were trained or designed for clinical notes and do not perform sufficiently well to build a public database of radiology reports. We developed and evaluated six ensemble models based on three publically available de-identification tools: MIT de-id, NeuroNER, and Philter. A set of 1023 reports was set aside as the testing partition. Two individuals with medical training annotated the test set for PHI; differences were resolved by consensus. Ensemble methods included simple voting schemes (1-Vote, 2-Votes, and 3-Votes), a decision tree, a naïve Bayesian classifier, and Adaboost boosting. The 1-Vote ensemble achieved recall of 998 / 1043 (95.7%); the 3-Votes ensemble had precision of 1035 / 1043 (99.2%). F1 scores were: 93.4% for the decision tree, 71.2% for the naïve Bayesian classifier, and 87.5% for the boosting method. Basic voting algorithms and machine learning classifiers incorporating the predictions of multiple tools can outperform each tool acting alone in de-identifying radiology reports. Ensemble methods hold substantial potential to improve automated de-identification tools for radiology reports to make such reports more available for research use to improve patient care and outcomes.
Collapse
Affiliation(s)
- Hannah Horng
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA, USA
| | - Jackson Steinkamp
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, USA
| | - Charles E Kahn
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, USA.
- Institute for Biomedical Informatics, University of Pennsylvania, Philadelphia, PA, USA.
| | - Tessa S Cook
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, USA
| |
Collapse
|
12
|
Hespel AM, Zhang Y, Basran PS. Artificial intelligence 101 for veterinary diagnostic imaging. Vet Radiol Ultrasound 2022; 63 Suppl 1:817-827. [PMID: 36514230 DOI: 10.1111/vru.13160] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2021] [Revised: 01/18/2022] [Accepted: 02/08/2022] [Indexed: 12/15/2022] Open
Abstract
The prevalence and pervasiveness of artificial intelligence (AI) with medical images in veterinary and human medicine is rapidly increasing. This article provides essential definitions of AI with medical images with a focus on veterinary radiology. Machine learning methods common in medical image analysis are compared, and a detailed description of convolutional neural networks commonly used in deep learning classification and regression models is provided. A brief introduction to natural language processing (NLP) and its utility in machine learning is also provided. NLP can economize the creation of "truth-data" needed when training AI systems for both diagnostic radiology and radiation oncology applications. The goal of this publication is to provide veterinarians, veterinary radiologists, and radiation oncologists the necessary background needed to understand and comprehend AI-focused research projects and publications.
Collapse
Affiliation(s)
- Adrien-Maxence Hespel
- Department of Small Animal Clinical Sciences, University of Tennessee, Knoxville, Tennessee, USA
| | - Youshan Zhang
- Department of Clinical Sciences, Cornell University, Ithaca, New York, USA
| | - Parminder S Basran
- Department of Clinical Sciences, Cornell University, Ithaca, New York, USA
| |
Collapse
|
13
|
Cho A, Min IK, Hong S, Chung HS, Lee HS, Kim JH. Effect of Applying a Real-Time Medical Record Input Assistance System With Voice Artificial Intelligence on Triage Task Performance in the Emergency Department: Prospective Interventional Study. JMIR Med Inform 2022; 10:e39892. [PMID: 36044254 PMCID: PMC9475416 DOI: 10.2196/39892] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2022] [Revised: 07/27/2022] [Accepted: 08/15/2022] [Indexed: 11/13/2022] Open
Abstract
BACKGROUND Natural language processing has been established as an important tool when using unstructured text data; however, most studies in the medical field have been limited to a retrospective analysis of text entered manually by humans. Little research has focused on applying natural language processing to the conversion of raw voice data generated in the clinical field into text using speech-to-text algorithms. OBJECTIVE In this study, we investigated the promptness and reliability of a real-time medical record input assistance system with voice artificial intelligence (RMIS-AI) and compared it to the manual method for triage tasks in the emergency department. METHODS From June 4, 2021, to September 12, 2021, RMIS-AI, using a machine learning engine trained with 1717 triage cases over 6 months, was prospectively applied in clinical practice in a triage unit. We analyzed a total of 1063 triage tasks performed by 19 triage nurses who agreed to participate. The primary outcome was the time for participants to perform the triage task. RESULTS The median time for participants to perform the triage task was 204 (IQR 155, 277) seconds by RMIS-AI and 231 (IQR 180, 313) seconds using manual method; this difference was statistically significant (P<.001). Most variables required for entry in the triage note showed a higher record completion rate by the manual method, but in the recording of additional chief concerns and past medical history, RMIS-AI showed a higher record completion rate than the manual method. Categorical variables entered by RMIS-AI showed less accuracy compared with continuous variables, such as vital signs. CONCLUSIONS RMIS-AI improves the promptness in performing triage tasks as compared to using the manual input method. However, to make it a reliable alternative to the conventional method, technical supplementation and additional research should be pursued.
Collapse
Affiliation(s)
- Ara Cho
- Department of Emergency Medicine, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - In Kyung Min
- Department of Research Affairs, Biostatistics Collaboration Unit, Yonsei University College, Seoul, Republic of Korea
| | - Seungkyun Hong
- CONNECT-AI Research Center, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Hyun Soo Chung
- Department of Emergency Medicine, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Hyun Sim Lee
- Department of Emergency Nursing, Yonsei University Health System, Seoul, Republic of Korea
| | - Ji Hoon Kim
- Department of Emergency Medicine, Yonsei University College of Medicine, Seoul, Republic of Korea
| |
Collapse
|
14
|
Linna N, Kahn CE. Applications of Natural Language Processing in Radiology: A Systematic Review. Int J Med Inform 2022; 163:104779. [DOI: 10.1016/j.ijmedinf.2022.104779] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Revised: 03/28/2022] [Accepted: 04/21/2022] [Indexed: 12/27/2022]
|
15
|
Basran PS, Appleby RB. The unmet potential of artificial intelligence in veterinary medicine. Am J Vet Res 2022; 83:385-392. [PMID: 35353711 DOI: 10.2460/ajvr.22.03.0038] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
Abstract
Veterinary medicine is a broad and growing discipline that includes topics such as companion animal health, population medicine and zoonotic diseases, and agriculture. In this article, we provide insight on how artificial intelligence works and how it is currently applied in veterinary medicine. We also discuss its potential in veterinary medicine. Given the rapid pace of research and commercial product developments in this area, the next several years will pose challenges to understanding, interpreting, and adopting this powerful and evolving technology. Artificial intelligence has the potential to enable veterinarians to perform tasks more efficiently while providing new insights for the management and treatment of disorders. It is our hope that this will translate to better quality of life for animals and those who care for them.
Collapse
Affiliation(s)
- Parminder S Basran
- Department of Clinical Sciences, College of Veterinary Medicine, Cornell University, Ithaca, NY
| | - Ryan B Appleby
- Department of Clinical Studies, Ontario Veterinary College, University of Guelph, Guelph, ON, Canada
| |
Collapse
|
16
|
Introduction to Artificial Intelligence in Medicine. Artif Intell Med 2022. [DOI: 10.1007/978-3-030-64573-1_27] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
17
|
Yin Z, Wong STC. Artificial intelligence unifies knowledge and actions in drug repositioning. Emerg Top Life Sci 2021; 5:803-813. [PMID: 34881780 PMCID: PMC8923082 DOI: 10.1042/etls20210223] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2021] [Revised: 11/08/2021] [Accepted: 11/09/2021] [Indexed: 11/17/2022]
Abstract
Drug repositioning aims to reuse existing drugs, shelved drugs, or drug candidates that failed clinical trials for other medical indications. Its attraction is sprung from the reduction in risk associated with safety testing of new medications and the time to get a known drug into the clinics. Artificial Intelligence (AI) has been recently pursued to speed up drug repositioning and discovery. The essence of AI in drug repositioning is to unify the knowledge and actions, i.e. incorporating real-world and experimental data to map out the best way forward to identify effective therapeutics against a disease. In this review, we share positive expectations for the evolution of AI and drug repositioning and summarize the role of AI in several methods of drug repositioning.
Collapse
Affiliation(s)
- Zheng Yin
- Department of Systems Medicine and Bioengineering, Houston Methodist Cancer Center and Ting Tsung & Wei Fong Chao Center for BRAIN, Houston Methodist Research Institute, Weill Cornell Medicine, Houston, TX 77030, U.S.A
| | - Stephen T C Wong
- Department of Systems Medicine and Bioengineering, Houston Methodist Cancer Center and Ting Tsung & Wei Fong Chao Center for BRAIN, Houston Methodist Research Institute, Weill Cornell Medicine, Houston, TX 77030, U.S.A
| |
Collapse
|
18
|
de Oliveira JM, da Costa CA, Antunes RS. Data structuring of electronic health records: a systematic review. HEALTH AND TECHNOLOGY 2021. [DOI: 10.1007/s12553-021-00607-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
|
19
|
Olthof AW, van Ooijen PMA, Cornelissen LJ. Deep Learning-Based Natural Language Processing in Radiology: The Impact of Report Complexity, Disease Prevalence, Dataset Size, and Algorithm Type on Model Performance. J Med Syst 2021; 45:91. [PMID: 34480231 PMCID: PMC8416876 DOI: 10.1007/s10916-021-01761-4] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2021] [Accepted: 08/04/2021] [Indexed: 12/12/2022]
Abstract
In radiology, natural language processing (NLP) allows the extraction of valuable information from radiology reports. It can be used for various downstream tasks such as quality improvement, epidemiological research, and monitoring guideline adherence. Class imbalance, variation in dataset size, variation in report complexity, and algorithm type all influence NLP performance but have not yet been systematically and interrelatedly evaluated. In this study, we investigate these factors on the performance of four types [a fully connected neural network (Dense), a long short-term memory recurrent neural network (LSTM), a convolutional neural network (CNN), and a Bidirectional Encoder Representations from Transformers (BERT)] of deep learning-based NLP. Two datasets consisting of radiologist-annotated reports of both trauma radiographs (n = 2469) and chest radiographs and computer tomography (CT) studies (n = 2255) were split into training sets (80%) and testing sets (20%). The training data was used as a source to train all four model types in 84 experiments (Fracture-data) and 45 experiments (Chest-data) with variation in size and prevalence. The performance was evaluated on sensitivity, specificity, positive predictive value, negative predictive value, area under the curve, and F score. After the NLP of radiology reports, all four model-architectures demonstrated high performance with metrics up to > 0.90. CNN, LSTM, and Dense were outperformed by the BERT algorithm because of its stable results despite variation in training size and prevalence. Awareness of variation in prevalence is warranted because it impacts sensitivity and specificity in opposite directions.
Collapse
Affiliation(s)
- A W Olthof
- Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Hanzeplein 1, Groningen, The Netherlands. .,Treant Health Care Group, Department of Radiology, Dr G.H. Amshoffweg 1, Hoogeveen, The Netherlands. .,Hospital Group Twente (ZGT), Department of Radiology, Almelo, The Netherlands.
| | - P M A van Ooijen
- Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Hanzeplein 1, Groningen, The Netherlands.,Data Science Center in Health (DASH), University of Groningen, University Medical Center Groningen, Machine Learning Lab, L.J, Zielstraweg 2, Groningen, The Netherlands
| | - L J Cornelissen
- Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Hanzeplein 1, Groningen, The Netherlands.,COSMONiO Imaging BV, L.J, Zielstraweg 2, Groningen, The Netherlands
| |
Collapse
|
20
|
Vaz JM, Balaji S. Convolutional neural networks (CNNs): concepts and applications in pharmacogenomics. Mol Divers 2021; 25:1569-1584. [PMID: 34031788 PMCID: PMC8342355 DOI: 10.1007/s11030-021-10225-3] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2021] [Accepted: 04/21/2021] [Indexed: 12/17/2022]
Abstract
Convolutional neural networks (CNNs) have been used to extract information from various datasets of different dimensions. This approach has led to accurate interpretations in several subfields of biological research, like pharmacogenomics, addressing issues previously faced by other computational methods. With the rising attention for personalized and precision medicine, scientists and clinicians have now turned to artificial intelligence systems to provide them with solutions for therapeutics development. CNNs have already provided valuable insights into biological data transformation. Due to the rise of interest in precision and personalized medicine, in this review, we have provided a brief overview of the possibilities of implementing CNNs as an effective tool for analyzing one-dimensional biological data, such as nucleotide and protein sequences, as well as small molecular data, e.g., simplified molecular-input line-entry specification, InChI, binary fingerprints, etc., to categorize the models based on their objective and also highlight various challenges. The review is organized into specific research domains that participate in pharmacogenomics for a more comprehensive understanding. Furthermore, the future intentions of deep learning are outlined.
Collapse
Affiliation(s)
- Joel Markus Vaz
- Department of Biotechnology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - S Balaji
- Department of Biotechnology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India.
| |
Collapse
|
21
|
ter Haar Romeny BM. Introduction to Artificial Intelligence in Medicine. Artif Intell Med 2021. [DOI: 10.1007/978-3-030-58080-3_27-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|